var/home/core/zuul-output/0000755000175000017500000000000015073400035014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073421061015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005562456315073421053017715 0ustar rootrootOct 14 07:51:05 crc systemd[1]: Starting Kubernetes Kubelet... Oct 14 07:51:05 crc restorecon[4717]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:05 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:51:06 crc restorecon[4717]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 14 07:51:07 crc kubenswrapper[5002]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.478377 5002 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482063 5002 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482079 5002 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482085 5002 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482089 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482093 5002 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482097 5002 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482100 5002 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482111 5002 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482117 5002 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482122 5002 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482125 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482129 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482133 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482137 5002 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482140 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482143 5002 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482147 5002 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482150 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482154 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482157 5002 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482160 5002 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482164 5002 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482167 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482172 5002 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482176 5002 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482180 5002 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482184 5002 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482188 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482191 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482196 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482199 5002 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482203 5002 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482207 5002 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482211 5002 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482215 5002 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482218 5002 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482222 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482226 5002 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482230 5002 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482235 5002 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482240 5002 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482243 5002 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482247 5002 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482250 5002 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482254 5002 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482257 5002 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482261 5002 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482264 5002 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482267 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482271 5002 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482274 5002 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482277 5002 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482281 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482284 5002 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482287 5002 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482291 5002 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482294 5002 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482298 5002 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482301 5002 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482304 5002 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482309 5002 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482313 5002 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482318 5002 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482322 5002 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482327 5002 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482331 5002 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482334 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482338 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482341 5002 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482345 5002 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.482349 5002 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483299 5002 flags.go:64] FLAG: --address="0.0.0.0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483313 5002 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483323 5002 flags.go:64] FLAG: --anonymous-auth="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483328 5002 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483334 5002 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483339 5002 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483345 5002 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483350 5002 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483355 5002 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483359 5002 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483364 5002 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483368 5002 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483372 5002 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483376 5002 flags.go:64] FLAG: --cgroup-root="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483380 5002 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483384 5002 flags.go:64] FLAG: --client-ca-file="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483388 5002 flags.go:64] FLAG: --cloud-config="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483392 5002 flags.go:64] FLAG: --cloud-provider="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483396 5002 flags.go:64] FLAG: --cluster-dns="[]" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483400 5002 flags.go:64] FLAG: --cluster-domain="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483404 5002 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483409 5002 flags.go:64] FLAG: --config-dir="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483412 5002 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483417 5002 flags.go:64] FLAG: --container-log-max-files="5" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483423 5002 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483427 5002 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483432 5002 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483436 5002 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483440 5002 flags.go:64] FLAG: --contention-profiling="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483444 5002 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483449 5002 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483454 5002 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483458 5002 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483463 5002 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483467 5002 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483471 5002 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483476 5002 flags.go:64] FLAG: --enable-load-reader="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483480 5002 flags.go:64] FLAG: --enable-server="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483484 5002 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483490 5002 flags.go:64] FLAG: --event-burst="100" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483495 5002 flags.go:64] FLAG: --event-qps="50" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483499 5002 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483503 5002 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483507 5002 flags.go:64] FLAG: --eviction-hard="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483512 5002 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483516 5002 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483520 5002 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483524 5002 flags.go:64] FLAG: --eviction-soft="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483529 5002 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483532 5002 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483536 5002 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483540 5002 flags.go:64] FLAG: --experimental-mounter-path="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483544 5002 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483548 5002 flags.go:64] FLAG: --fail-swap-on="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483552 5002 flags.go:64] FLAG: --feature-gates="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483558 5002 flags.go:64] FLAG: --file-check-frequency="20s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483562 5002 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483566 5002 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483570 5002 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483575 5002 flags.go:64] FLAG: --healthz-port="10248" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483579 5002 flags.go:64] FLAG: --help="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483582 5002 flags.go:64] FLAG: --hostname-override="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483586 5002 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483590 5002 flags.go:64] FLAG: --http-check-frequency="20s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483594 5002 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483598 5002 flags.go:64] FLAG: --image-credential-provider-config="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483602 5002 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483606 5002 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483610 5002 flags.go:64] FLAG: --image-service-endpoint="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483614 5002 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483618 5002 flags.go:64] FLAG: --kube-api-burst="100" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483622 5002 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483627 5002 flags.go:64] FLAG: --kube-api-qps="50" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483631 5002 flags.go:64] FLAG: --kube-reserved="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483635 5002 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483638 5002 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483643 5002 flags.go:64] FLAG: --kubelet-cgroups="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483646 5002 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483650 5002 flags.go:64] FLAG: --lock-file="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483655 5002 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483659 5002 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483663 5002 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483669 5002 flags.go:64] FLAG: --log-json-split-stream="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483673 5002 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483677 5002 flags.go:64] FLAG: --log-text-split-stream="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483681 5002 flags.go:64] FLAG: --logging-format="text" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483685 5002 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483689 5002 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483693 5002 flags.go:64] FLAG: --manifest-url="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483697 5002 flags.go:64] FLAG: --manifest-url-header="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483703 5002 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483707 5002 flags.go:64] FLAG: --max-open-files="1000000" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483712 5002 flags.go:64] FLAG: --max-pods="110" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483716 5002 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483720 5002 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483724 5002 flags.go:64] FLAG: --memory-manager-policy="None" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483729 5002 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483733 5002 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483737 5002 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483741 5002 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483750 5002 flags.go:64] FLAG: --node-status-max-images="50" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483754 5002 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483758 5002 flags.go:64] FLAG: --oom-score-adj="-999" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483762 5002 flags.go:64] FLAG: --pod-cidr="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483767 5002 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483773 5002 flags.go:64] FLAG: --pod-manifest-path="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483777 5002 flags.go:64] FLAG: --pod-max-pids="-1" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483781 5002 flags.go:64] FLAG: --pods-per-core="0" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483786 5002 flags.go:64] FLAG: --port="10250" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483790 5002 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483794 5002 flags.go:64] FLAG: --provider-id="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483798 5002 flags.go:64] FLAG: --qos-reserved="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483802 5002 flags.go:64] FLAG: --read-only-port="10255" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483806 5002 flags.go:64] FLAG: --register-node="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483810 5002 flags.go:64] FLAG: --register-schedulable="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483814 5002 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483821 5002 flags.go:64] FLAG: --registry-burst="10" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483825 5002 flags.go:64] FLAG: --registry-qps="5" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483829 5002 flags.go:64] FLAG: --reserved-cpus="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483845 5002 flags.go:64] FLAG: --reserved-memory="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483851 5002 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483854 5002 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483858 5002 flags.go:64] FLAG: --rotate-certificates="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483863 5002 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483867 5002 flags.go:64] FLAG: --runonce="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483871 5002 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483875 5002 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483880 5002 flags.go:64] FLAG: --seccomp-default="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483886 5002 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483893 5002 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483903 5002 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483914 5002 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483925 5002 flags.go:64] FLAG: --storage-driver-password="root" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483935 5002 flags.go:64] FLAG: --storage-driver-secure="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483946 5002 flags.go:64] FLAG: --storage-driver-table="stats" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483955 5002 flags.go:64] FLAG: --storage-driver-user="root" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483965 5002 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483984 5002 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.483993 5002 flags.go:64] FLAG: --system-cgroups="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484003 5002 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484018 5002 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484027 5002 flags.go:64] FLAG: --tls-cert-file="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484036 5002 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484047 5002 flags.go:64] FLAG: --tls-min-version="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484057 5002 flags.go:64] FLAG: --tls-private-key-file="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484067 5002 flags.go:64] FLAG: --topology-manager-policy="none" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484077 5002 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484086 5002 flags.go:64] FLAG: --topology-manager-scope="container" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484096 5002 flags.go:64] FLAG: --v="2" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484108 5002 flags.go:64] FLAG: --version="false" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484121 5002 flags.go:64] FLAG: --vmodule="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484143 5002 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.484153 5002 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484367 5002 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484379 5002 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484389 5002 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484399 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484407 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484415 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484429 5002 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484438 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484447 5002 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484456 5002 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484465 5002 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484473 5002 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484481 5002 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484490 5002 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484499 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484507 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484518 5002 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484526 5002 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484535 5002 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484543 5002 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484552 5002 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484560 5002 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484568 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484576 5002 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484585 5002 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484596 5002 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484607 5002 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484617 5002 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484626 5002 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484638 5002 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484648 5002 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484657 5002 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484666 5002 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484675 5002 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484685 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484694 5002 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484705 5002 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484716 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484734 5002 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484744 5002 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484753 5002 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484762 5002 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484771 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484779 5002 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484788 5002 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484796 5002 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484805 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484814 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484824 5002 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484857 5002 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484866 5002 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484878 5002 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484889 5002 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484899 5002 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484908 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484917 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484925 5002 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484934 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484943 5002 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484951 5002 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484959 5002 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484968 5002 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484976 5002 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484985 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.484993 5002 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485002 5002 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485010 5002 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485019 5002 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485027 5002 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485035 5002 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.485046 5002 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.485060 5002 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.496305 5002 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.496353 5002 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496429 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496441 5002 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496447 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496452 5002 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496457 5002 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496462 5002 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496466 5002 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496470 5002 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496475 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496479 5002 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496484 5002 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496488 5002 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496494 5002 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496501 5002 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496510 5002 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496516 5002 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496521 5002 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496526 5002 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496530 5002 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496535 5002 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496539 5002 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496544 5002 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496548 5002 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496554 5002 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496559 5002 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496563 5002 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496568 5002 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496573 5002 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496578 5002 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496582 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496586 5002 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496591 5002 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496596 5002 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496600 5002 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496604 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496608 5002 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496611 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496614 5002 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496619 5002 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496624 5002 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496628 5002 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496632 5002 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496636 5002 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496640 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496644 5002 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496648 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496652 5002 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496657 5002 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496661 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496665 5002 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496669 5002 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496672 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496676 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496680 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496683 5002 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496687 5002 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496690 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496693 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496697 5002 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496702 5002 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496706 5002 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496709 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496712 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496716 5002 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496720 5002 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496723 5002 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496727 5002 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496732 5002 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496738 5002 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496743 5002 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496748 5002 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.496756 5002 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496898 5002 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496908 5002 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496915 5002 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496919 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496924 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496928 5002 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496933 5002 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496939 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496943 5002 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496948 5002 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496953 5002 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496958 5002 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496963 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496967 5002 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496971 5002 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496974 5002 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496980 5002 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496985 5002 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496989 5002 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496993 5002 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.496998 5002 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497002 5002 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497007 5002 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497012 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497017 5002 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497022 5002 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497027 5002 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497032 5002 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497037 5002 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497042 5002 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497046 5002 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497051 5002 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497055 5002 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497060 5002 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497065 5002 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497070 5002 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497074 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497100 5002 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497104 5002 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497110 5002 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497116 5002 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497121 5002 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497126 5002 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497131 5002 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497136 5002 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497141 5002 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497145 5002 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497167 5002 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497172 5002 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497177 5002 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497182 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497186 5002 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497191 5002 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497197 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497203 5002 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497208 5002 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497214 5002 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497220 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497225 5002 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497230 5002 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497235 5002 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497240 5002 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497246 5002 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497250 5002 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497257 5002 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497262 5002 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497267 5002 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497272 5002 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497278 5002 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497283 5002 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.497288 5002 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.497296 5002 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.497499 5002 server.go:940] "Client rotation is on, will bootstrap in background" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.501969 5002 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.502331 5002 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.504046 5002 server.go:997] "Starting client certificate rotation" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.504072 5002 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.504213 5002 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-07 09:21:05.528033126 +0000 UTC Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.504257 5002 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 577h29m58.023777993s for next certificate rotation Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.532933 5002 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.537341 5002 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.555915 5002 log.go:25] "Validated CRI v1 runtime API" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.594907 5002 log.go:25] "Validated CRI v1 image API" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.597139 5002 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.602421 5002 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-14-07-46-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.602470 5002 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.632907 5002 manager.go:217] Machine: {Timestamp:2025-10-14 07:51:07.630025493 +0000 UTC m=+0.611265025 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a0ba6967-5ef5-4816-9f6e-d1367009bebc BootID:8e326842-e02f-4877-83ac-d4b3808edbb7 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:63:c1:c4 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:63:c1:c4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f2:ea:ba Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:77:34:e0 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:33:2c:53 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c5:e6:f1 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:f8:b1:cb Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:8a:1c:7a:82:a3 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f2:2e:d2:65:3d:f3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.633380 5002 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.633592 5002 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.634379 5002 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.634695 5002 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.634740 5002 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.635099 5002 topology_manager.go:138] "Creating topology manager with none policy" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.635119 5002 container_manager_linux.go:303] "Creating device plugin manager" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.635895 5002 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.635956 5002 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.636252 5002 state_mem.go:36] "Initialized new in-memory state store" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.636386 5002 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.640367 5002 kubelet.go:418] "Attempting to sync node with API server" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.640436 5002 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.640475 5002 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.640495 5002 kubelet.go:324] "Adding apiserver pod source" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.640514 5002 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.648497 5002 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.649823 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.649862 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.649940 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.649960 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.651098 5002 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.652632 5002 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654490 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654534 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654551 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654564 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654585 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654598 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654611 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654640 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654660 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654677 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654704 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.654719 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.656083 5002 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.656773 5002 server.go:1280] "Started kubelet" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.657177 5002 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.657867 5002 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.658531 5002 server.go:460] "Adding debug handlers to kubelet server" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.658914 5002 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.659007 5002 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:07 crc systemd[1]: Started Kubernetes Kubelet. Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.659922 5002 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.659997 5002 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.660223 5002 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 15:21:01.821423107 +0000 UTC Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.660276 5002 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 727h29m54.161151276s for next certificate rotation Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.660510 5002 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.660589 5002 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.660614 5002 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.665946 5002 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.667161 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="200ms" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.668710 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.668807 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.667903 5002 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.74:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e4c2eaa3c306d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-14 07:51:07.656720493 +0000 UTC m=+0.637959975,LastTimestamp:2025-10-14 07:51:07.656720493 +0000 UTC m=+0.637959975,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.671376 5002 factory.go:55] Registering systemd factory Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.671412 5002 factory.go:221] Registration of the systemd container factory successfully Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.673092 5002 factory.go:153] Registering CRI-O factory Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.673141 5002 factory.go:221] Registration of the crio container factory successfully Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.673285 5002 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.673351 5002 factory.go:103] Registering Raw factory Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.673386 5002 manager.go:1196] Started watching for new ooms in manager Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.674482 5002 manager.go:319] Starting recovery of all containers Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678218 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678270 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678283 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678298 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678312 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678325 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678338 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678351 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678365 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678377 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678390 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678403 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678415 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678430 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678445 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678497 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678530 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678545 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678557 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678570 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678584 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678597 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678609 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678621 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678632 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678645 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678682 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678695 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678707 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678719 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678732 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678745 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678757 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678771 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678783 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678797 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678809 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678822 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678850 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678886 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678900 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678913 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678927 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678941 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678954 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678969 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678981 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.678995 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679008 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679023 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679035 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679049 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679067 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679079 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679092 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679106 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679121 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679133 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679145 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679158 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679169 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679182 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679193 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679205 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679216 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679228 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679240 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679252 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679264 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679276 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679288 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679301 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679312 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679324 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679335 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679347 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679359 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679372 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679383 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679396 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679408 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679421 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679432 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679444 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679455 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.679468 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681428 5002 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681456 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681469 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681482 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681494 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681507 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681518 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681529 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681540 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681552 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681563 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681576 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681587 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681597 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681609 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681620 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681632 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681725 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681739 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681760 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681773 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681786 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681800 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681813 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681826 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681860 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681873 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681932 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681947 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681959 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681970 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681982 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.681996 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682022 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682033 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682045 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682059 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682070 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682081 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682092 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682101 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682112 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682124 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682135 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682147 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682158 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682171 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682184 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682197 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682210 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682222 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682236 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682250 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682262 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682275 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682287 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682298 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682311 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682327 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682338 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682350 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682364 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682374 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682387 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682401 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682414 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682427 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682439 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682451 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682464 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682477 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682490 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682502 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682513 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682528 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682539 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682551 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682562 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682574 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682585 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682598 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682610 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682625 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682635 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682647 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682658 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682670 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682681 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682693 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682705 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682720 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682732 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682743 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682755 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682767 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682779 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682791 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682805 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682818 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682846 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682858 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682870 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682882 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682896 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682909 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682921 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682933 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682946 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682958 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682969 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682982 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.682993 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683006 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683019 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683033 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683046 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683059 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683072 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683085 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683097 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683111 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683124 5002 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683136 5002 reconstruct.go:97] "Volume reconstruction finished" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.683144 5002 reconciler.go:26] "Reconciler: start to sync state" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.707497 5002 manager.go:324] Recovery completed Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.717232 5002 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.717320 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719045 5002 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719105 5002 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719141 5002 kubelet.go:2335] "Starting kubelet main sync loop" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.719230 5002 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.719761 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.721244 5002 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.721265 5002 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.721307 5002 state_mem.go:36] "Initialized new in-memory state store" Oct 14 07:51:07 crc kubenswrapper[5002]: W1014 07:51:07.721318 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.721446 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.742427 5002 policy_none.go:49] "None policy: Start" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.743302 5002 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.743327 5002 state_mem.go:35] "Initializing new in-memory state store" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.761421 5002 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.799381 5002 manager.go:334] "Starting Device Plugin manager" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.799471 5002 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.799497 5002 server.go:79] "Starting device plugin registration server" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.800229 5002 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.800308 5002 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.800608 5002 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.800767 5002 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.800784 5002 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.809757 5002 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.819643 5002 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.820028 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.822353 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.822385 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.822394 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.822524 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.823382 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.823403 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.823412 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.823963 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.824009 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.824032 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.824056 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.823980 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825138 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825162 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825172 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825802 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825862 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825868 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825875 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.825923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826158 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826282 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826313 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826950 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.826996 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827108 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827167 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827213 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827237 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827541 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827582 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827861 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.827874 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828057 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828089 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828677 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828893 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828926 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.828937 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.868045 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="400ms" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885587 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885622 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885648 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885672 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885693 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885712 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885733 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885777 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885818 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885866 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885881 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885893 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885907 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885921 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.885934 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.900643 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.901795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.901861 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.901882 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.901918 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:07 crc kubenswrapper[5002]: E1014 07:51:07.902368 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986777 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986825 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986867 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986888 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986907 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986925 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986948 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986958 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987018 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986913 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987025 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987081 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986968 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987020 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987053 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.986967 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987224 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987257 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987285 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987313 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987322 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987361 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987366 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987410 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987482 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987505 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987528 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:07 crc kubenswrapper[5002]: I1014 07:51:07.987383 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.102528 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.104137 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.104195 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.104213 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.104248 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:08 crc kubenswrapper[5002]: E1014 07:51:08.104760 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.158461 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.177124 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.199466 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.216416 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.221670 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.228008 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7f78f2e76fec315715bc57e32aec496fb4b26a99765c844bd846fd9d7d371c25 WatchSource:0}: Error finding container 7f78f2e76fec315715bc57e32aec496fb4b26a99765c844bd846fd9d7d371c25: Status 404 returned error can't find the container with id 7f78f2e76fec315715bc57e32aec496fb4b26a99765c844bd846fd9d7d371c25 Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.229158 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a07ad131617282fcef68c571d5eee0940f4959e17b5eb106334eef9d053e0595 WatchSource:0}: Error finding container a07ad131617282fcef68c571d5eee0940f4959e17b5eb106334eef9d053e0595: Status 404 returned error can't find the container with id a07ad131617282fcef68c571d5eee0940f4959e17b5eb106334eef9d053e0595 Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.245205 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4e5dce72137ce204e9551d50dd1ee259531b18df17178422779d2467623d5f20 WatchSource:0}: Error finding container 4e5dce72137ce204e9551d50dd1ee259531b18df17178422779d2467623d5f20: Status 404 returned error can't find the container with id 4e5dce72137ce204e9551d50dd1ee259531b18df17178422779d2467623d5f20 Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.251466 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6c10b5f917170032554954a1877998aaee5e1dba5d1e17cb0c5eb01a4b534494 WatchSource:0}: Error finding container 6c10b5f917170032554954a1877998aaee5e1dba5d1e17cb0c5eb01a4b534494: Status 404 returned error can't find the container with id 6c10b5f917170032554954a1877998aaee5e1dba5d1e17cb0c5eb01a4b534494 Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.252885 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b68b5c61d60b8add751f6a4bf109bbe3bce3722547caa735bfd33d7c4dc8cef8 WatchSource:0}: Error finding container b68b5c61d60b8add751f6a4bf109bbe3bce3722547caa735bfd33d7c4dc8cef8: Status 404 returned error can't find the container with id b68b5c61d60b8add751f6a4bf109bbe3bce3722547caa735bfd33d7c4dc8cef8 Oct 14 07:51:08 crc kubenswrapper[5002]: E1014 07:51:08.269769 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="800ms" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.505275 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.508219 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.508284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.508298 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.508327 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:08 crc kubenswrapper[5002]: E1014 07:51:08.508759 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.660353 5002 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.724877 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4e5dce72137ce204e9551d50dd1ee259531b18df17178422779d2467623d5f20"} Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.726104 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a07ad131617282fcef68c571d5eee0940f4959e17b5eb106334eef9d053e0595"} Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.727537 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7f78f2e76fec315715bc57e32aec496fb4b26a99765c844bd846fd9d7d371c25"} Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.728704 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b68b5c61d60b8add751f6a4bf109bbe3bce3722547caa735bfd33d7c4dc8cef8"} Oct 14 07:51:08 crc kubenswrapper[5002]: I1014 07:51:08.729948 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6c10b5f917170032554954a1877998aaee5e1dba5d1e17cb0c5eb01a4b534494"} Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.845383 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:08 crc kubenswrapper[5002]: E1014 07:51:08.845496 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:08 crc kubenswrapper[5002]: W1014 07:51:08.947495 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:08 crc kubenswrapper[5002]: E1014 07:51:08.947653 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:09 crc kubenswrapper[5002]: E1014 07:51:09.070923 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="1.6s" Oct 14 07:51:09 crc kubenswrapper[5002]: W1014 07:51:09.079610 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:09 crc kubenswrapper[5002]: E1014 07:51:09.079710 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:09 crc kubenswrapper[5002]: W1014 07:51:09.225362 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:09 crc kubenswrapper[5002]: E1014 07:51:09.225507 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.308897 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.310269 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.310344 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.310361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.310429 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:09 crc kubenswrapper[5002]: E1014 07:51:09.311096 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.660402 5002 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.734950 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.735001 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.735014 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.735024 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.735192 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736563 5002 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630" exitCode=0 Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736624 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736642 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736650 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736670 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.736739 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738529 5002 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17" exitCode=0 Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738581 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738604 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738647 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.738659 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.739473 5002 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955" exitCode=0 Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.739550 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.739577 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.743348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.743382 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.743391 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.746274 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.746314 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.746327 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.749253 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b" exitCode=0 Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.749304 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b"} Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.749664 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.752242 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.752290 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.752304 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.759400 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.760543 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.760587 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:09 crc kubenswrapper[5002]: I1014 07:51:09.760598 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: W1014 07:51:10.493777 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:10 crc kubenswrapper[5002]: E1014 07:51:10.494241 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:10 crc kubenswrapper[5002]: W1014 07:51:10.579998 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:10 crc kubenswrapper[5002]: E1014 07:51:10.580115 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.660754 5002 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Oct 14 07:51:10 crc kubenswrapper[5002]: E1014 07:51:10.672542 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="3.2s" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.743324 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.754434 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.754491 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.754506 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.754552 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.756439 5002 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7" exitCode=0 Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.756508 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.756578 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.757336 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.757361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.757371 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.759446 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.759470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.759480 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.759537 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.760786 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.760807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.760815 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.762098 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ef8eefae32a67ba6fcd11ac7ad8d4d5a29f3c4197290614149e780404c2fcccb"} Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.762144 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.762152 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763340 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763368 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763380 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763365 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.763420 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.913306 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.916546 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.916580 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.916590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.916615 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:10 crc kubenswrapper[5002]: E1014 07:51:10.917131 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Oct 14 07:51:10 crc kubenswrapper[5002]: I1014 07:51:10.977296 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.121676 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.132380 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.769602 5002 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845" exitCode=0 Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.770111 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.770166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845"} Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.771675 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.771723 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.771742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.777257 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a"} Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.777351 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.777410 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.777451 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.777741 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.778612 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779608 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779906 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779926 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779671 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779549 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.780045 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.780060 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.780076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.780082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.779865 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:11 crc kubenswrapper[5002]: I1014 07:51:11.983045 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792300 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f"} Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792424 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8"} Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792450 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb"} Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792332 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792471 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792549 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.792476 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794705 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794734 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794771 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794933 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.794953 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:12 crc kubenswrapper[5002]: I1014 07:51:12.929386 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800661 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d"} Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800720 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800742 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec"} Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800781 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800800 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.800978 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802234 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802244 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802241 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802382 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.802404 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.803088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.803121 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:13 crc kubenswrapper[5002]: I1014 07:51:13.803129 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.118412 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.120667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.120731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.120742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.120770 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.802582 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.803736 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.803790 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.803805 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.842588 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.842726 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.842764 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.843972 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.844005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.844016 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.983793 5002 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 07:51:14 crc kubenswrapper[5002]: I1014 07:51:14.983974 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:51:15 crc kubenswrapper[5002]: I1014 07:51:15.471948 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 14 07:51:15 crc kubenswrapper[5002]: I1014 07:51:15.806775 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:15 crc kubenswrapper[5002]: I1014 07:51:15.808090 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:15 crc kubenswrapper[5002]: I1014 07:51:15.808156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:15 crc kubenswrapper[5002]: I1014 07:51:15.808180 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:16 crc kubenswrapper[5002]: I1014 07:51:16.654649 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 14 07:51:16 crc kubenswrapper[5002]: I1014 07:51:16.808509 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:16 crc kubenswrapper[5002]: I1014 07:51:16.809495 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:16 crc kubenswrapper[5002]: I1014 07:51:16.809530 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:16 crc kubenswrapper[5002]: I1014 07:51:16.809540 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:17 crc kubenswrapper[5002]: E1014 07:51:17.809922 5002 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 07:51:17 crc kubenswrapper[5002]: I1014 07:51:17.887186 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:17 crc kubenswrapper[5002]: I1014 07:51:17.887430 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:17 crc kubenswrapper[5002]: I1014 07:51:17.888995 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:17 crc kubenswrapper[5002]: I1014 07:51:17.889052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:17 crc kubenswrapper[5002]: I1014 07:51:17.889132 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:20 crc kubenswrapper[5002]: I1014 07:51:20.751186 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:20 crc kubenswrapper[5002]: I1014 07:51:20.751424 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:20 crc kubenswrapper[5002]: I1014 07:51:20.753026 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:20 crc kubenswrapper[5002]: I1014 07:51:20.753086 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:20 crc kubenswrapper[5002]: I1014 07:51:20.753103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:21 crc kubenswrapper[5002]: W1014 07:51:21.307478 5002 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.307581 5002 trace.go:236] Trace[2117270608]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:51:11.305) (total time: 10002ms): Oct 14 07:51:21 crc kubenswrapper[5002]: Trace[2117270608]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:51:21.307) Oct 14 07:51:21 crc kubenswrapper[5002]: Trace[2117270608]: [10.002080468s] [10.002080468s] END Oct 14 07:51:21 crc kubenswrapper[5002]: E1014 07:51:21.307612 5002 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.661263 5002 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.939801 5002 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.939945 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.958908 5002 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 07:51:21 crc kubenswrapper[5002]: I1014 07:51:21.958989 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 07:51:22 crc kubenswrapper[5002]: I1014 07:51:22.945596 5002 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]log ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]etcd ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/generic-apiserver-start-informers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/priority-and-fairness-filter ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-apiextensions-informers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-apiextensions-controllers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/crd-informer-synced ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-system-namespaces-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/bootstrap-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/start-kube-aggregator-informers ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-registration-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-discovery-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]autoregister-completion ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-openapi-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 14 07:51:22 crc kubenswrapper[5002]: livez check failed Oct 14 07:51:22 crc kubenswrapper[5002]: I1014 07:51:22.945695 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:51:24 crc kubenswrapper[5002]: I1014 07:51:24.983333 5002 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 07:51:24 crc kubenswrapper[5002]: I1014 07:51:24.983414 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.407456 5002 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.661276 5002 apiserver.go:52] "Watching apiserver" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.673110 5002 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.673630 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.674274 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.674792 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.674885 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.674968 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:25 crc kubenswrapper[5002]: E1014 07:51:25.675109 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:25 crc kubenswrapper[5002]: E1014 07:51:25.675181 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.677660 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678022 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678413 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678768 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678799 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678883 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:25 crc kubenswrapper[5002]: E1014 07:51:25.679020 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678910 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678799 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678921 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.678780 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.681453 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.710105 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.728006 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.747229 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.765265 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.766823 5002 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.782629 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.796050 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:25 crc kubenswrapper[5002]: I1014 07:51:25.810761 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.691068 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.713382 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.714491 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.717112 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.728126 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.742994 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.758673 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.775325 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.789491 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.806164 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.821002 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.836093 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.863312 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.878652 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.893250 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.907935 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:26 crc kubenswrapper[5002]: E1014 07:51:26.926767 5002 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.929424 5002 trace.go:236] Trace[1592773079]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:51:15.382) (total time: 11546ms): Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1592773079]: ---"Objects listed" error: 11546ms (07:51:26.929) Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1592773079]: [11.546932982s] [11.546932982s] END Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.929467 5002 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.929584 5002 trace.go:236] Trace[1632155658]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:51:16.041) (total time: 10888ms): Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1632155658]: ---"Objects listed" error: 10888ms (07:51:26.929) Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1632155658]: [10.888402272s] [10.888402272s] END Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.929618 5002 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.931195 5002 trace.go:236] Trace[1417764536]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:51:11.964) (total time: 14966ms): Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1417764536]: ---"Objects listed" error: 14965ms (07:51:26.930) Oct 14 07:51:26 crc kubenswrapper[5002]: Trace[1417764536]: [14.966241878s] [14.966241878s] END Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.931255 5002 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 14 07:51:26 crc kubenswrapper[5002]: I1014 07:51:26.931206 5002 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 14 07:51:26 crc kubenswrapper[5002]: E1014 07:51:26.931425 5002 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031629 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031665 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031684 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031713 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031732 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031753 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031768 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031782 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031798 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031813 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031827 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031869 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031888 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031929 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031960 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031976 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.031995 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032031 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032068 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032101 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032119 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032135 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032152 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032171 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032188 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032208 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032228 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032206 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032245 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032340 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032368 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032391 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032411 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032429 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032448 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032464 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032487 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032506 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032524 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032544 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032565 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032581 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032601 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032619 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032609 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032628 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032639 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032747 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032792 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032833 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032896 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032898 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032929 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032937 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032955 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.032982 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033002 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033017 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033086 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033114 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033142 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033150 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033167 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033194 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033219 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033242 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033267 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033296 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033324 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033353 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033381 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033406 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033428 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033450 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033473 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033495 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033521 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033545 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033570 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033596 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033619 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033642 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033664 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033692 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033714 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033736 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033758 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033781 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033802 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033823 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033865 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033887 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033908 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033929 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033952 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033978 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034000 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034024 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034047 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034072 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034099 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034122 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034193 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034217 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034238 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034268 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034292 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034316 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034339 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034363 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034386 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034410 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034432 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034453 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034477 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034506 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034529 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034552 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034575 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034597 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034620 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034643 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034665 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034687 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034711 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034732 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034754 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034777 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034798 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034821 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034861 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034893 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034916 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034939 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034962 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034986 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035531 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035571 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035613 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035642 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035667 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035705 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035800 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035853 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035930 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035962 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035985 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036009 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036032 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036067 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036100 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036134 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036165 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036189 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036212 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036243 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036273 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036299 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036350 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036376 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036402 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036430 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036454 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037125 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037166 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037190 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037214 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037238 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037263 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037293 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037318 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037341 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037370 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037395 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037422 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037446 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037469 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037494 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037519 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037546 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037569 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037592 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037616 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037638 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037661 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037685 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037720 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037753 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037790 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037822 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037868 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037895 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037918 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037946 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037976 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038001 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038026 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038053 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038081 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038115 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038154 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038187 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038248 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038297 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038344 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038392 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038445 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038496 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038551 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038592 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038646 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038696 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039151 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039215 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039255 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039295 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039372 5002 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039397 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039416 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039435 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039456 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039474 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039493 5002 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039520 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033167 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033339 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039825 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033360 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033474 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033662 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033770 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033774 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033846 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.033880 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034022 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034121 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034173 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034241 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.034291 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035029 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035101 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035216 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035300 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035371 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035423 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035572 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035755 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035825 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036189 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036241 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.035741 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036407 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036438 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036440 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036447 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036510 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036591 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036675 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.040332 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036696 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.036994 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037006 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037071 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037722 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038021 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.037967 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038257 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038142 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038736 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.038410 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039140 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039311 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039339 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039409 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039484 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039613 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.039791 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.040296 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041060 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041090 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041216 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041389 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041622 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041669 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041667 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041655 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041712 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041732 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.041749 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.042308 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.042780 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.042328 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.042886 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043020 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043196 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043210 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043353 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043421 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043548 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043668 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.043794 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.044227 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.044808 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.044819 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.044471 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.045001 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.045297 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.045505 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.045662 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.045611 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.045759 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:27.545730446 +0000 UTC m=+20.526969908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.045795 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.046371 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.046528 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.046718 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.046740 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:27.546716102 +0000 UTC m=+20.527955554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.046739 5002 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.046971 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047104 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047227 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047242 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047447 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047598 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047713 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047765 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.047771 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.048111 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.048195 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.048461 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.048795 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.049111 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.049473 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.049709 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.049925 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.050590 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051071 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051238 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051450 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051585 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051705 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051957 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.053978 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.054412 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.054414 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.050468 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.054863 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.054927 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.050222 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.051733 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.055363 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.055631 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.056039 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.056216 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:27.554956616 +0000 UTC m=+20.536196068 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.056373 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.057018 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.057364 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.057932 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.057948 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058213 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058219 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058359 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058441 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.056415 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058967 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.058979 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.059108 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.059294 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.059406 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.059531 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.059850 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.060488 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.061191 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.061274 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.062829 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.062878 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.063778 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.064493 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.064603 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.064984 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.065086 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.066435 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.066513 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.067710 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.068080 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.068381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.068697 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.069158 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.069673 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071282 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071639 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071683 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071708 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071826 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071944 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072309 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072467 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072493 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072513 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072530 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072547 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072574 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072588 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072607 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072624 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072638 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.072673 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.072710 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.072722 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072742 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072950 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.072973 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.073061 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:27.573038509 +0000 UTC m=+20.554277951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.076553 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.076574 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.071735 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.077430 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.077573 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.077553 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.078293 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.078349 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.078451 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:27.57844046 +0000 UTC m=+20.559679912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.078481 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.078644 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.078666 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.082801 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.083085 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.083256 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.089195 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.095497 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.096601 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.100193 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.102424 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.140371 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.140424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.140479 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.140681 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141065 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141125 5002 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141142 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141158 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141171 5002 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141184 5002 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141196 5002 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141209 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141222 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141234 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141247 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141259 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141271 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141283 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141295 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141307 5002 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141318 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141330 5002 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141342 5002 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141354 5002 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141380 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141392 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141404 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141416 5002 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141428 5002 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141439 5002 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141452 5002 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141464 5002 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141476 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141489 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141505 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141517 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141531 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141545 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141560 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141573 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141584 5002 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141596 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141610 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141701 5002 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141721 5002 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141737 5002 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141755 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141771 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141784 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141796 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141809 5002 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141823 5002 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141851 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141863 5002 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141875 5002 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141886 5002 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141898 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141909 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141921 5002 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141932 5002 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141942 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141956 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141984 5002 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.141995 5002 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142011 5002 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142175 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142189 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142203 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142216 5002 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142227 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142238 5002 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142250 5002 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142260 5002 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142271 5002 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142283 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142294 5002 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142305 5002 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142316 5002 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142326 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142340 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142351 5002 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142362 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142399 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142412 5002 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142423 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142437 5002 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142449 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142460 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142503 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142514 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142525 5002 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142568 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142586 5002 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142603 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142624 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142640 5002 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142658 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142676 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142692 5002 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142709 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142726 5002 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142743 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142759 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142787 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142806 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142822 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142865 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142882 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142910 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142930 5002 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142947 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142966 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.142984 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143002 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143022 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143041 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143059 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143080 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143097 5002 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143114 5002 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143130 5002 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143146 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143237 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143261 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143281 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143300 5002 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143327 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143343 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143360 5002 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143377 5002 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143394 5002 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143412 5002 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143429 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143446 5002 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143463 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143480 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143496 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143513 5002 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143528 5002 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143544 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143563 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143582 5002 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143599 5002 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143615 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143633 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143655 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143687 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143710 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143731 5002 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143768 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143792 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143813 5002 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143867 5002 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143885 5002 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.143992 5002 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144015 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144032 5002 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144049 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144065 5002 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144083 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144130 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144397 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144424 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144442 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144460 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144638 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144670 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144694 5002 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144710 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144727 5002 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144744 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144802 5002 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144823 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144866 5002 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144884 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144902 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144922 5002 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144939 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144958 5002 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144975 5002 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.144992 5002 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145009 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145027 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145044 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145061 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145077 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145097 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145114 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145132 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.145151 5002 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.199431 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.216536 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.230183 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:51:27 crc kubenswrapper[5002]: W1014 07:51:27.230496 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-9f403c2d14ce35a147f721ccad99e22a7fa503e7043b8281f2ce0faa1ccda95b WatchSource:0}: Error finding container 9f403c2d14ce35a147f721ccad99e22a7fa503e7043b8281f2ce0faa1ccda95b: Status 404 returned error can't find the container with id 9f403c2d14ce35a147f721ccad99e22a7fa503e7043b8281f2ce0faa1ccda95b Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.547248 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.547360 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.547438 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:28.547411926 +0000 UTC m=+21.528651378 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.547471 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.547533 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:28.547516859 +0000 UTC m=+21.528756361 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.648452 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.648513 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.648540 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648645 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648670 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648692 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648704 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648733 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:28.648707983 +0000 UTC m=+21.629947435 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648764 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648800 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648811 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:28.648800097 +0000 UTC m=+21.630039549 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648818 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.648953 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:28.64892148 +0000 UTC m=+21.630160972 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.720116 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.720271 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.720427 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.720677 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.721200 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:27 crc kubenswrapper[5002]: E1014 07:51:27.721309 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.726049 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.727101 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.728946 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.729961 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.731324 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.732125 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.733079 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.734727 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.736065 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.737312 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.738388 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.739721 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.740742 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.741794 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.742893 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.744637 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.745326 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.745741 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.746362 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.747000 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.747521 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.748337 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.750580 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.751305 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.751700 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.752370 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.753054 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.753555 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.754150 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.755553 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.755920 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.756391 5002 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.756494 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.758479 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.758968 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.759353 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.761002 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.762071 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.762551 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.763507 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.764134 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.765150 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.765764 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.766904 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.767472 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.768302 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.768809 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.769739 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.770481 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.771464 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.771987 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.772814 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.773331 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.773887 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.774681 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.775075 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.787012 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.797096 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.812200 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.825446 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.840893 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.841238 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"06b9680519e9b01553b74a9262d5017ddbbd1f016e67a654ca5c04f493661b4b"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.842911 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.844227 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.848064 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a" exitCode=255 Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.848354 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.852890 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d02d562665ea19f48f279072babd5e315ba6dccfde4bfb6e1d1ae7d299c65a60"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.858405 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.858480 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.858494 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9f403c2d14ce35a147f721ccad99e22a7fa503e7043b8281f2ce0faa1ccda95b"} Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.884811 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.889501 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.890053 5002 scope.go:117] "RemoveContainer" containerID="2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.928341 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.946280 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.949606 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.972275 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:27 crc kubenswrapper[5002]: I1014 07:51:27.984950 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.000305 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.013502 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.025347 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.040506 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.055951 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.089271 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.107002 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.121481 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.137555 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.150563 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.558160 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.558310 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.558391 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:30.558345206 +0000 UTC m=+23.539584658 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.558528 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.558650 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:30.558622203 +0000 UTC m=+23.539861695 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.659603 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.659689 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.659728 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.659757 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.659926 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:30.65989045 +0000 UTC m=+23.641129942 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.659940 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.659976 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.659995 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.660005 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.660060 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.660078 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.660078 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:30.660055844 +0000 UTC m=+23.641295336 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:28 crc kubenswrapper[5002]: E1014 07:51:28.660166 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:30.660141596 +0000 UTC m=+23.641381038 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.862333 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.864136 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746"} Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.869348 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.881540 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.900493 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.916085 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.936118 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.952330 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.977542 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:28 crc kubenswrapper[5002]: I1014 07:51:28.994067 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.006789 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.030127 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.052407 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.076360 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.093612 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.106475 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.118786 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.129068 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.138726 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.719727 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.719864 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.719733 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:29 crc kubenswrapper[5002]: E1014 07:51:29.719965 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:29 crc kubenswrapper[5002]: E1014 07:51:29.720050 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:29 crc kubenswrapper[5002]: E1014 07:51:29.720185 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.868475 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783"} Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.868862 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.893382 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.913498 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.933386 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.948280 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.978233 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:29 crc kubenswrapper[5002]: I1014 07:51:29.996188 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.019142 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.039594 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.576103 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.576263 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.576357 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:34.576331702 +0000 UTC m=+27.557571184 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.576465 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.576567 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:34.576546048 +0000 UTC m=+27.557785510 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.677589 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.677682 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:30 crc kubenswrapper[5002]: I1014 07:51:30.677755 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.677906 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.677995 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:34.677970589 +0000 UTC m=+27.659210081 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678006 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678067 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678095 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678017 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678172 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678191 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678195 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:34.678160503 +0000 UTC m=+27.659400015 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:30 crc kubenswrapper[5002]: E1014 07:51:30.678238 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:34.678224115 +0000 UTC m=+27.659463577 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:31 crc kubenswrapper[5002]: I1014 07:51:31.719902 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:31 crc kubenswrapper[5002]: I1014 07:51:31.719996 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:31 crc kubenswrapper[5002]: E1014 07:51:31.720358 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:31 crc kubenswrapper[5002]: I1014 07:51:31.720008 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:31 crc kubenswrapper[5002]: E1014 07:51:31.720520 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:31 crc kubenswrapper[5002]: E1014 07:51:31.720647 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:31 crc kubenswrapper[5002]: I1014 07:51:31.989453 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:31 crc kubenswrapper[5002]: I1014 07:51:31.994331 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.001822 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.020721 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.041995 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.061915 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.084966 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.103031 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.118239 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.137242 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.151917 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.168797 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.185694 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.203348 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.220905 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.238523 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.253555 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.272414 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.306506 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:32 crc kubenswrapper[5002]: I1014 07:51:32.323914 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.331797 5002 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.334215 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.334255 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.334267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.334335 5002 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.345238 5002 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.345541 5002 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.347112 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.347531 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.347716 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.347954 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.348112 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.380937 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.385529 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.385573 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.385588 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.385610 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.385627 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.404155 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.412414 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.412446 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.412455 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.412468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.412477 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.427764 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.431626 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.431791 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.431952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.432085 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.432210 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.452396 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.457539 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.457732 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.457860 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.457988 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.458077 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.474801 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.474979 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.477077 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.477112 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.477124 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.477140 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.477152 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.580268 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.580666 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.580807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.580988 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.581118 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.683801 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.683861 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.683872 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.683886 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.683896 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.722123 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.722234 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.722486 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.722535 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.722566 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:33 crc kubenswrapper[5002]: E1014 07:51:33.722603 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.786512 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.786558 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.786569 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.786585 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.786596 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.888759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.889028 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.889124 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.889405 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.889431 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.991819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.991890 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.991901 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.991915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:33 crc kubenswrapper[5002]: I1014 07:51:33.991942 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:33Z","lastTransitionTime":"2025-10-14T07:51:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.094047 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.094087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.094096 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.094111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.094122 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.161670 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-q57t4"] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.162201 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.163658 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.164150 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.164473 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.188884 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.196786 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.196813 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.196854 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.196875 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.196884 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.203921 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.210051 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2wsj\" (UniqueName: \"kubernetes.io/projected/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-kube-api-access-x2wsj\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.210245 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-hosts-file\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.214702 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.225885 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.234008 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.244467 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.255865 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.292177 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.299170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.299216 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.299229 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.299246 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.299259 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.311230 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2wsj\" (UniqueName: \"kubernetes.io/projected/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-kube-api-access-x2wsj\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.311284 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-hosts-file\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.311381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-hosts-file\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.319951 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.339140 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2wsj\" (UniqueName: \"kubernetes.io/projected/db1bd93d-cd11-4b20-a5c5-1758b2a53b3c-kube-api-access-x2wsj\") pod \"node-resolver-q57t4\" (UID: \"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\") " pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.390105 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.401757 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.402008 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.402087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.402157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.402223 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.474262 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-q57t4" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.504418 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.504446 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.504454 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.504467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.504477 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.568506 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7btf2"] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.568734 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-s4jgg"] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.569110 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-9n4rz"] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.569295 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.569548 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.569956 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.571270 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.571496 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.571704 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.572725 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.573026 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.573228 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574134 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574233 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574505 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574647 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574529 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.574596 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.588787 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.601751 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.607303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.607339 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.607363 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.607378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.607387 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.614938 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615054 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-socket-dir-parent\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.615080 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:42.615054438 +0000 UTC m=+35.596293890 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615106 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-kubelet\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615131 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-hostroot\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615150 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615168 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f3de631-7681-4103-8510-2b5545bbe8cd-rootfs\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615224 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615297 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj84k\" (UniqueName: \"kubernetes.io/projected/6f3de631-7681-4103-8510-2b5545bbe8cd-kube-api-access-zj84k\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615316 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-conf-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615333 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-etc-kubernetes\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.615341 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.615406 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:42.615388806 +0000 UTC m=+35.596628258 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615355 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-system-cni-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615452 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-netns\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615472 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-os-release\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615493 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615522 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwqz\" (UniqueName: \"kubernetes.io/projected/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-kube-api-access-wpwqz\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615539 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615584 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-system-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615611 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cnibin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615654 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f3de631-7681-4103-8510-2b5545bbe8cd-proxy-tls\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615673 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cni-binary-copy\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615715 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615743 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-os-release\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615774 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-k8s-cni-cncf-io\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615801 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-bin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615826 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmkwj\" (UniqueName: \"kubernetes.io/projected/bf44e87d-acc3-427a-b2ce-24daf814d1c6-kube-api-access-nmkwj\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615862 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f3de631-7681-4103-8510-2b5545bbe8cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615894 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-daemon-config\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615908 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-multus-certs\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615931 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-multus\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.615948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cnibin\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.618497 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.631446 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.641681 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.661689 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.673293 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.683939 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.693015 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.702910 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.709659 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.709687 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.709697 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.709712 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.709748 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.713317 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716503 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-system-cni-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716533 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-netns\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716557 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-os-release\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716582 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716604 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-system-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716604 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-system-cni-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716673 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-netns\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716706 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716623 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716742 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-os-release\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716755 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwqz\" (UniqueName: \"kubernetes.io/projected/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-kube-api-access-wpwqz\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716780 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cnibin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716808 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f3de631-7681-4103-8510-2b5545bbe8cd-proxy-tls\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716829 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cni-binary-copy\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716879 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716902 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716924 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-os-release\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716945 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-k8s-cni-cncf-io\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716966 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f3de631-7681-4103-8510-2b5545bbe8cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.716991 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717021 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-bin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717042 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmkwj\" (UniqueName: \"kubernetes.io/projected/bf44e87d-acc3-427a-b2ce-24daf814d1c6-kube-api-access-nmkwj\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717066 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717086 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-daemon-config\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717108 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-multus-certs\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717128 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cnibin\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717152 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-multus\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717173 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-kubelet\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717197 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-hostroot\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717217 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717238 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f3de631-7681-4103-8510-2b5545bbe8cd-rootfs\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717261 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-socket-dir-parent\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717298 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-conf-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717340 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-etc-kubernetes\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717370 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj84k\" (UniqueName: \"kubernetes.io/projected/6f3de631-7681-4103-8510-2b5545bbe8cd-kube-api-access-zj84k\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717489 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.717615 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-system-cni-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.717702 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.717749 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:42.717733691 +0000 UTC m=+35.698973153 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718047 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718070 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718068 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-bin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718082 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718125 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:42.718111921 +0000 UTC m=+35.699351483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718158 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-cni-multus\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718193 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-multus-certs\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718223 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cnibin\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718255 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-var-lib-kubelet\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718304 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-os-release\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718462 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-host-run-k8s-cni-cncf-io\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718537 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-conf-dir\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718539 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bf44e87d-acc3-427a-b2ce-24daf814d1c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718588 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6f3de631-7681-4103-8510-2b5545bbe8cd-rootfs\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718616 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-socket-dir-parent\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718631 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-hostroot\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718539 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bf44e87d-acc3-427a-b2ce-24daf814d1c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718650 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718673 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718682 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-etc-kubernetes\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718683 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-multus-daemon-config\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718691 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cnibin\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718686 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:34 crc kubenswrapper[5002]: E1014 07:51:34.718812 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:42.718799959 +0000 UTC m=+35.700039421 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.718887 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-cni-binary-copy\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.721248 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6f3de631-7681-4103-8510-2b5545bbe8cd-mcd-auth-proxy-config\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.721573 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6f3de631-7681-4103-8510-2b5545bbe8cd-proxy-tls\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.730099 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.735462 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj84k\" (UniqueName: \"kubernetes.io/projected/6f3de631-7681-4103-8510-2b5545bbe8cd-kube-api-access-zj84k\") pod \"machine-config-daemon-7btf2\" (UID: \"6f3de631-7681-4103-8510-2b5545bbe8cd\") " pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.735996 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmkwj\" (UniqueName: \"kubernetes.io/projected/bf44e87d-acc3-427a-b2ce-24daf814d1c6-kube-api-access-nmkwj\") pod \"multus-additional-cni-plugins-s4jgg\" (UID: \"bf44e87d-acc3-427a-b2ce-24daf814d1c6\") " pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.737414 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwqz\" (UniqueName: \"kubernetes.io/projected/359daa94-9198-48cf-bbea-a7d8cdb8f3c7-kube-api-access-wpwqz\") pod \"multus-9n4rz\" (UID: \"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\") " pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.744019 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.762141 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.776692 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.788969 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.807606 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.811571 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.811595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.811604 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.811619 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.811629 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.818993 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.827884 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.845588 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.858932 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.870148 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.878669 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.881407 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-q57t4" event={"ID":"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c","Type":"ContainerStarted","Data":"37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.881439 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-q57t4" event={"ID":"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c","Type":"ContainerStarted","Data":"0149e5a528fe5663c8aac2cafafba4d6b38f53ef55d930ad799e2e13ead2afc5"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.888742 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9n4rz" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.895177 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.898995 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.901222 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" Oct 14 07:51:34 crc kubenswrapper[5002]: W1014 07:51:34.909885 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f3de631_7681_4103_8510_2b5545bbe8cd.slice/crio-6478008cc26e8b4dafe653924b2103f943052940f68460f20115640a206a4662 WatchSource:0}: Error finding container 6478008cc26e8b4dafe653924b2103f943052940f68460f20115640a206a4662: Status 404 returned error can't find the container with id 6478008cc26e8b4dafe653924b2103f943052940f68460f20115640a206a4662 Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.914614 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.914731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.914756 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.914784 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.914807 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:34Z","lastTransitionTime":"2025-10-14T07:51:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.919339 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: W1014 07:51:34.929965 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf44e87d_acc3_427a_b2ce_24daf814d1c6.slice/crio-01e82e141aa3a889bcdca5ff53b2581716ae0f73f4c8f1f85e54d695ff055288 WatchSource:0}: Error finding container 01e82e141aa3a889bcdca5ff53b2581716ae0f73f4c8f1f85e54d695ff055288: Status 404 returned error can't find the container with id 01e82e141aa3a889bcdca5ff53b2581716ae0f73f4c8f1f85e54d695ff055288 Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.936970 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.948302 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gd5gm"] Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.951383 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.956491 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.956686 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.956828 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.957240 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.957613 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.957958 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.958342 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.968335 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:34 crc kubenswrapper[5002]: I1014 07:51:34.984527 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.002896 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016013 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016884 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016948 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.016980 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018770 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018803 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018821 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018855 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96bqb\" (UniqueName: \"kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018873 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.018992 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019032 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019064 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019191 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019246 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019271 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019291 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019326 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019344 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019366 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019401 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019431 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019460 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019501 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.019524 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.028175 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.046935 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.060051 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.080622 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.094603 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.107993 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120026 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120189 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120200 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120219 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120232 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120445 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120472 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120493 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120515 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96bqb\" (UniqueName: \"kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120540 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120567 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120727 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120740 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120767 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120787 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120823 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120862 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120878 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120904 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120922 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120943 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120967 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.120989 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121007 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121020 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121035 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121094 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121124 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121148 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121169 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121192 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121214 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121308 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121420 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121457 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121485 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121507 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121718 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121969 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.121988 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.122027 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.122208 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.123791 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.124912 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.130732 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.137603 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96bqb\" (UniqueName: \"kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb\") pod \"ovnkube-node-gd5gm\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.151018 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.165469 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.177033 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.220879 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.222830 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.222880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.222892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.222910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.222920 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.247012 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.257112 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.271039 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.282013 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.292066 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.310345 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.324753 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.324793 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.324803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.324818 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.324829 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.326359 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.339266 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.341788 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.349694 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: W1014 07:51:35.352939 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod858331e7_9f04_4a5d_8bf9_8307dfa68556.slice/crio-596453ff2a17a47d84491528bb31fc130ab25c0dbf9ee71db4b931d4af77ffbc WatchSource:0}: Error finding container 596453ff2a17a47d84491528bb31fc130ab25c0dbf9ee71db4b931d4af77ffbc: Status 404 returned error can't find the container with id 596453ff2a17a47d84491528bb31fc130ab25c0dbf9ee71db4b931d4af77ffbc Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.365024 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.384525 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.427143 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.427190 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.427206 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.427232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.427244 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.529127 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.529170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.529179 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.529195 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.529205 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.631521 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.631568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.631579 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.631595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.631606 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.719466 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.719491 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.719466 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:35 crc kubenswrapper[5002]: E1014 07:51:35.719602 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:35 crc kubenswrapper[5002]: E1014 07:51:35.719656 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:35 crc kubenswrapper[5002]: E1014 07:51:35.719792 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.733533 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.733599 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.733617 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.733641 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.733674 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.836343 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.836412 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.836429 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.836454 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.836471 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.886380 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8" exitCode=0 Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.886468 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.886504 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerStarted","Data":"01e82e141aa3a889bcdca5ff53b2581716ae0f73f4c8f1f85e54d695ff055288"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.888735 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.888779 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.888795 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"6478008cc26e8b4dafe653924b2103f943052940f68460f20115640a206a4662"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.890478 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerStarted","Data":"4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.890553 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerStarted","Data":"c7fbd2486e0c922a141c63efbaeae2f56e407fb8e9594c6e614bc3bfef972d18"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.892188 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" exitCode=0 Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.892236 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.892287 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"596453ff2a17a47d84491528bb31fc130ab25c0dbf9ee71db4b931d4af77ffbc"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.907372 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.924395 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.939425 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.939468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.939483 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.939507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.939523 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:35Z","lastTransitionTime":"2025-10-14T07:51:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.943437 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.960490 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:35 crc kubenswrapper[5002]: I1014 07:51:35.982526 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.012796 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.027570 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.040059 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.042678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.042703 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.042719 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.042736 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.042748 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.060072 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.080194 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.091364 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.105641 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.116132 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.133666 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.145232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.145259 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.145268 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.145281 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.145290 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.149762 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.161352 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.172991 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.188302 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.198464 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.209382 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.219685 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.230714 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.241939 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.247429 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.247453 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.247462 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.247475 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.247484 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.254497 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.268224 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.278715 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.317298 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.350104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.350958 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.351049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.351133 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.351233 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.353365 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.454498 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.454677 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.454823 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.454956 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.455058 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.557448 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.557561 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.557581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.557607 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.557625 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.659734 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.660245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.660266 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.660290 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.660307 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.763872 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.764018 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.764111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.764184 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.764261 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.867727 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.867764 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.867775 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.867789 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.867799 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.903434 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerStarted","Data":"1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.907320 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.907366 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.907379 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.907391 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.907402 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.920183 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.933270 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.945091 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.955169 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.968814 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.970245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.970502 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.970510 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.970523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.970533 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:36Z","lastTransitionTime":"2025-10-14T07:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:36 crc kubenswrapper[5002]: I1014 07:51:36.992702 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.017753 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.041818 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.057217 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.069683 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.072111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.072148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.072158 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.072170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.072180 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.078965 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.102726 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.116411 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.127674 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.175443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.175717 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.175882 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.176015 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.176131 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.278588 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.278922 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.279085 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.279187 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.279266 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.382034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.382909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.382990 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.383055 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.383117 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.485334 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.485378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.485389 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.485407 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.485418 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.587433 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.587467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.587477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.587493 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.587506 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.690087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.690139 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.690155 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.690174 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.690187 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.719792 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.719940 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.719966 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:37 crc kubenswrapper[5002]: E1014 07:51:37.720090 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:37 crc kubenswrapper[5002]: E1014 07:51:37.720225 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:37 crc kubenswrapper[5002]: E1014 07:51:37.720406 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.734545 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.746045 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.757355 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.769706 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.786483 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.793355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.793394 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.793406 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.793424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.793436 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.809662 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.826750 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.839909 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.859830 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.875296 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.886510 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.895877 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.895931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.895947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.895969 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.895984 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:37Z","lastTransitionTime":"2025-10-14T07:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.903354 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.912056 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b" exitCode=0 Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.912115 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.917083 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.922549 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.950235 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.961915 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.981645 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:37 crc kubenswrapper[5002]: I1014 07:51:37.993722 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005283 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005293 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005306 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005314 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.005800 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.018535 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.038677 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.053622 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.066011 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.083096 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.093746 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.104396 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.108572 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.108597 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.108606 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.108619 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.108627 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.117081 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.129643 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.147339 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.210913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.210964 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.210977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.210997 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.211010 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.313667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.313704 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.313712 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.313726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.313734 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.334998 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-lb9sb"] Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.335481 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.338460 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.338721 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.338733 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.338878 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.352241 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.367655 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.394983 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.407773 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.415677 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.415755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.415777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.415806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.415895 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.423256 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.437691 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.451728 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.458274 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sks28\" (UniqueName: \"kubernetes.io/projected/14694d10-47a7-493e-8973-feb8372782ca-kube-api-access-sks28\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.458377 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/14694d10-47a7-493e-8973-feb8372782ca-host\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.458495 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/14694d10-47a7-493e-8973-feb8372782ca-serviceca\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.467287 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.498424 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.512422 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.518424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.518552 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.518579 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.518605 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.518624 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.556685 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.559441 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sks28\" (UniqueName: \"kubernetes.io/projected/14694d10-47a7-493e-8973-feb8372782ca-kube-api-access-sks28\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.559531 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/14694d10-47a7-493e-8973-feb8372782ca-host\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.559624 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/14694d10-47a7-493e-8973-feb8372782ca-serviceca\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.559749 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/14694d10-47a7-493e-8973-feb8372782ca-host\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.561825 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/14694d10-47a7-493e-8973-feb8372782ca-serviceca\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.604904 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sks28\" (UniqueName: \"kubernetes.io/projected/14694d10-47a7-493e-8973-feb8372782ca-kube-api-access-sks28\") pod \"node-ca-lb9sb\" (UID: \"14694d10-47a7-493e-8973-feb8372782ca\") " pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.617206 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.622087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.622151 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.622169 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.622195 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.622213 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.647205 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lb9sb" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.664957 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: W1014 07:51:38.668893 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14694d10_47a7_493e_8973_feb8372782ca.slice/crio-526819ffc3e6a3c77d8b4b0257f581451449a977c25dcfb26a8a6bbd5cb20be1 WatchSource:0}: Error finding container 526819ffc3e6a3c77d8b4b0257f581451449a977c25dcfb26a8a6bbd5cb20be1: Status 404 returned error can't find the container with id 526819ffc3e6a3c77d8b4b0257f581451449a977c25dcfb26a8a6bbd5cb20be1 Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.695124 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.726413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.726467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.726485 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.726503 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.726517 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.739198 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.834213 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.834271 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.834288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.834310 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.834326 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.922328 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb" exitCode=0 Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.922398 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.923679 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lb9sb" event={"ID":"14694d10-47a7-493e-8973-feb8372782ca","Type":"ContainerStarted","Data":"526819ffc3e6a3c77d8b4b0257f581451449a977c25dcfb26a8a6bbd5cb20be1"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.933374 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.936474 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.936522 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.936534 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.936554 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.936567 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:38Z","lastTransitionTime":"2025-10-14T07:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.947649 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.964073 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.977235 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:38 crc kubenswrapper[5002]: I1014 07:51:38.991773 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.019744 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.032369 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.038300 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.038337 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.038348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.038362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.038373 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.051546 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.097414 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.137324 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.140911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.140975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.140986 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.141021 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.141031 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.174953 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.211170 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.249149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.249181 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.249190 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.249203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.249210 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.257853 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.297320 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.333934 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.352259 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.352327 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.352350 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.352377 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.352400 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.455376 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.455443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.455465 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.455492 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.455513 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.558828 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.558932 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.558968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.558998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.559020 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.662592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.662656 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.662674 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.662697 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.662716 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.720220 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.720241 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:39 crc kubenswrapper[5002]: E1014 07:51:39.720424 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.720472 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:39 crc kubenswrapper[5002]: E1014 07:51:39.720608 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:39 crc kubenswrapper[5002]: E1014 07:51:39.720799 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.765734 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.765793 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.765812 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.765865 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.765885 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.869642 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.869724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.869749 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.869781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.869805 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.932861 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246" exitCode=0 Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.932984 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.936478 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lb9sb" event={"ID":"14694d10-47a7-493e-8973-feb8372782ca","Type":"ContainerStarted","Data":"71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.943284 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.957160 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.972695 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.972770 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.972793 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.972826 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.972895 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:39Z","lastTransitionTime":"2025-10-14T07:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:39 crc kubenswrapper[5002]: I1014 07:51:39.976773 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.005958 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.030325 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.046818 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.065465 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.081760 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.082043 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.082913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.083005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.083034 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.086927 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.103772 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.131478 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.151920 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.164371 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.175263 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.185430 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.185577 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.185654 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.185731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.185811 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.188480 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.205764 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.218070 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.236058 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.256154 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.269209 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.282361 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.288905 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.289034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.289116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.289212 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.289308 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.297445 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.319983 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.339536 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.355663 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.370777 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.386401 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.391911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.391966 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.391981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.391998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.392010 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.408481 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.442093 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.462765 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.495720 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.495814 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.495834 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.496653 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.496687 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.500953 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.537331 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.599652 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.599707 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.599724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.599747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.599764 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.702680 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.702778 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.702797 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.702821 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.702859 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.805921 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.805985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.806003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.806071 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.806089 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.908946 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.909000 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.909016 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.909039 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.909057 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:40Z","lastTransitionTime":"2025-10-14T07:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.952211 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerStarted","Data":"b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4"} Oct 14 07:51:40 crc kubenswrapper[5002]: I1014 07:51:40.990311 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.011927 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.012909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.012992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.013011 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.013069 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.013091 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.032522 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.050014 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.072255 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.090822 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.107610 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.116007 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.116080 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.116101 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.116539 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.116598 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.128928 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.148204 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.177611 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.198609 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.212699 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.219006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.219033 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.219044 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.219062 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.219074 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.225805 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.243914 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.259258 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.321281 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.321369 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.321387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.321413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.321433 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.424274 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.424322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.424342 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.424372 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.424396 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.529923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.529992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.530067 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.530141 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.530182 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.633567 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.633631 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.633655 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.633687 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.633709 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.720124 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.720156 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.720223 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:41 crc kubenswrapper[5002]: E1014 07:51:41.720299 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:41 crc kubenswrapper[5002]: E1014 07:51:41.720390 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:41 crc kubenswrapper[5002]: E1014 07:51:41.720510 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.737179 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.737253 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.737279 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.737313 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.737337 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.840597 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.840731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.840749 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.840773 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.840789 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.943652 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.943704 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.943721 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.943745 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.943762 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:41Z","lastTransitionTime":"2025-10-14T07:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.960946 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4" exitCode=0 Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.961055 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.971480 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876"} Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.972277 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.972320 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:41 crc kubenswrapper[5002]: I1014 07:51:41.992503 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.011353 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.014178 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.019213 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.037113 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.047400 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.047448 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.047464 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.047488 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.047506 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.052792 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.072368 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.090517 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.105193 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.122283 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.140344 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.150409 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.150464 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.150482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.150505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.150524 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.166765 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.188634 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.206742 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.224265 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.240285 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.252914 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.253133 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.253290 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.253422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.253606 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.260438 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.286119 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.302767 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.320118 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.335784 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356535 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356614 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356627 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.356946 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.384183 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.403110 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.422246 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.437130 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.454628 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.458651 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.458712 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.458731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.458757 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.458774 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.472624 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.486874 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.505452 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.523472 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.559232 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.561182 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.561252 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.561271 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.561297 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.561316 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.617051 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.617294 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.617350 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:51:58.617311314 +0000 UTC m=+51.598550806 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.617475 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.617589 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:58.61755677 +0000 UTC m=+51.598796282 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.663934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.663973 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.663983 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.663995 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.664004 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.718302 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.718414 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718529 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718569 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718591 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718593 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718681 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:58.718653153 +0000 UTC m=+51.699892635 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.718714 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:58.718700814 +0000 UTC m=+51.699940306 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.766954 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.767017 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.767033 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.767057 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.767079 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.818895 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.819182 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.819219 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.819240 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:42 crc kubenswrapper[5002]: E1014 07:51:42.819352 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:58.819301073 +0000 UTC m=+51.800540565 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.869293 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.869345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.869361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.869386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.869402 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.972987 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.973038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.973054 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.973076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.973093 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:42Z","lastTransitionTime":"2025-10-14T07:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.979903 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf44e87d-acc3-427a-b2ce-24daf814d1c6" containerID="cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c" exitCode=0 Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.980012 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerDied","Data":"cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c"} Oct 14 07:51:42 crc kubenswrapper[5002]: I1014 07:51:42.980122 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:42.999962 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.025542 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.044302 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.063034 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.076193 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.076240 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.076258 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.076282 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.076300 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.083734 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.107173 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.127168 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.142766 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.154358 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.172506 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.179517 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.179577 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.179586 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.179600 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.179609 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.187482 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.200804 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.218882 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.235653 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.259771 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.282184 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.282208 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.282216 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.282227 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.282235 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.384773 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.384825 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.384869 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.384898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.384917 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.492022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.492084 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.492103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.492128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.492146 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.573635 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.573712 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.573740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.573774 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.573792 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.603887 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.608348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.608384 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.608395 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.608427 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.608440 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.624306 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.628640 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.628682 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.628695 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.628713 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.628724 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.644420 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.648207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.648233 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.648242 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.648258 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.648267 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.660494 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.666678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.666708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.666725 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.666739 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.666748 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.680083 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.680256 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.681609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.681647 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.681659 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.681676 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.681687 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.720194 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.720208 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.720215 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.720299 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.720424 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:43 crc kubenswrapper[5002]: E1014 07:51:43.720517 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.783615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.783665 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.783677 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.783710 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.783730 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.886220 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.886266 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.886278 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.886296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.886310 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.987599 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" event={"ID":"bf44e87d-acc3-427a-b2ce-24daf814d1c6","Type":"ContainerStarted","Data":"14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad"} Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.987723 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.988259 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.988304 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.988321 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.988343 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:43 crc kubenswrapper[5002]: I1014 07:51:43.988360 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:43Z","lastTransitionTime":"2025-10-14T07:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.011409 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.034338 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.091241 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.091278 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.091288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.091303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.091312 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.112228 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.129739 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.157565 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.170942 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.182585 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.194614 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.194637 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.194645 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.194658 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.194666 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.201674 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.223145 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.236988 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.248578 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.259019 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.278480 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.295821 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.296650 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.296707 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.296726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.296749 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.296766 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.307607 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:44Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.399509 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.399557 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.399574 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.399597 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.399614 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.503301 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.503362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.503379 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.503402 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.503419 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.607066 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.607118 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.607134 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.607160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.607177 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.710667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.710727 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.710752 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.710788 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.710824 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.815117 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.815170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.815186 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.815208 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.815225 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.917960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.918362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.918380 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.918402 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.918422 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:44Z","lastTransitionTime":"2025-10-14T07:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.993373 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/0.log" Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.997477 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876" exitCode=1 Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.997535 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876"} Oct 14 07:51:44 crc kubenswrapper[5002]: I1014 07:51:44.998698 5002 scope.go:117] "RemoveContainer" containerID="e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.020976 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.021029 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.021047 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.021075 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.021093 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.030197 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.049225 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.063898 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.082725 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.100657 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.123892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.123947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.123965 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.123991 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.124010 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.130439 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.152043 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.168135 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.185173 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.200660 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232129 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232827 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232894 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.232920 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.256865 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.277468 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.291511 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.301218 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.334777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.334819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.334849 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.334865 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.334877 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.437633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.437694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.437714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.437740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.437761 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.540714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.540807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.540877 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.540909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.540932 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.642777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.642821 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.642853 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.642871 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.642883 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.720122 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.720237 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:45 crc kubenswrapper[5002]: E1014 07:51:45.720287 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.720312 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:45 crc kubenswrapper[5002]: E1014 07:51:45.720484 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:45 crc kubenswrapper[5002]: E1014 07:51:45.720595 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.744360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.744401 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.744409 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.744422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.744431 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.846547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.846598 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.846613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.846633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.846648 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.949229 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.949290 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.949307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.949331 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:45 crc kubenswrapper[5002]: I1014 07:51:45.949348 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:45Z","lastTransitionTime":"2025-10-14T07:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.002348 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/0.log" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.005261 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.005379 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.018420 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.037136 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.052425 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.052545 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.052619 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.052662 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.052675 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.058519 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.079429 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.094643 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.115321 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.131558 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.155462 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.155513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.155544 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.155567 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.155582 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.164439 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.181558 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.197219 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.213549 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.236138 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.260125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.260470 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.260581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.260693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.260789 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.261552 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.276786 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.292056 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:46Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.364179 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.364232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.364245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.364263 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.364274 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.466952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.467368 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.467557 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.467768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.467973 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.570789 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.570892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.570915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.570944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.570997 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.674078 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.674919 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.675168 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.675347 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.675491 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.778992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.779422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.779639 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.779810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.780041 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.884068 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.884130 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.884147 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.884168 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.884186 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.987257 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.987595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.987736 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.987911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:46 crc kubenswrapper[5002]: I1014 07:51:46.988046 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:46Z","lastTransitionTime":"2025-10-14T07:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.012358 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/1.log" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.013464 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/0.log" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.018636 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87" exitCode=1 Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.018703 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.018785 5002 scope.go:117] "RemoveContainer" containerID="e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.020144 5002 scope.go:117] "RemoveContainer" containerID="f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87" Oct 14 07:51:47 crc kubenswrapper[5002]: E1014 07:51:47.020467 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.040140 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.061710 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4"] Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.062286 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.064894 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.065118 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.065247 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.085910 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.090892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.090957 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.090971 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.090991 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.091008 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.099969 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.112489 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.115937 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.153778 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.170888 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43d2f41c-af58-4a2a-840a-634408e94c04-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.171072 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.171126 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqq27\" (UniqueName: \"kubernetes.io/projected/43d2f41c-af58-4a2a-840a-634408e94c04-kube-api-access-cqq27\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.171240 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.173704 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.189109 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.194017 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.194063 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.194076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.194093 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.194105 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.202518 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.223935 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.241728 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.257815 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.272682 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43d2f41c-af58-4a2a-840a-634408e94c04-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.272743 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.272779 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqq27\" (UniqueName: \"kubernetes.io/projected/43d2f41c-af58-4a2a-840a-634408e94c04-kube-api-access-cqq27\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.272829 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.273904 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.274354 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43d2f41c-af58-4a2a-840a-634408e94c04-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.274620 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.280997 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43d2f41c-af58-4a2a-840a-634408e94c04-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.294676 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqq27\" (UniqueName: \"kubernetes.io/projected/43d2f41c-af58-4a2a-840a-634408e94c04-kube-api-access-cqq27\") pod \"ovnkube-control-plane-749d76644c-cpjn4\" (UID: \"43d2f41c-af58-4a2a-840a-634408e94c04\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.296690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.296754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.296781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.296812 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.296834 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.297801 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.326962 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.341659 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.359619 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.374959 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.388345 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.399436 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.402393 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.402435 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.402452 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.402476 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.402495 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: W1014 07:51:47.406439 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43d2f41c_af58_4a2a_840a_634408e94c04.slice/crio-65f5dd769fd3b003b9b61b326875a2b65691bd57540cd8a44f54e41c318e64bb WatchSource:0}: Error finding container 65f5dd769fd3b003b9b61b326875a2b65691bd57540cd8a44f54e41c318e64bb: Status 404 returned error can't find the container with id 65f5dd769fd3b003b9b61b326875a2b65691bd57540cd8a44f54e41c318e64bb Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.421208 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.433434 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.448955 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.459262 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.492998 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.505390 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.505429 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.505443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.505464 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.505479 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.510316 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.527467 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.540198 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.555813 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.579104 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.591996 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.606102 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.607304 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.607348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.607357 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.607370 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.607379 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.710694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.710732 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.710743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.710760 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.710771 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.720620 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.720815 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.721039 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:47 crc kubenswrapper[5002]: E1014 07:51:47.721030 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:47 crc kubenswrapper[5002]: E1014 07:51:47.721221 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:47 crc kubenswrapper[5002]: E1014 07:51:47.721396 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.739830 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.753756 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.766103 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.781497 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.796596 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.812317 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.812558 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.812644 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.812759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.812868 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.823787 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.843683 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.857672 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.872428 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.887640 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.892064 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.903645 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.915230 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.915274 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.915288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.915307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.915318 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:47Z","lastTransitionTime":"2025-10-14T07:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.924822 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.951077 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.961588 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.972022 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.980025 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:47 crc kubenswrapper[5002]: I1014 07:51:47.994791 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.006206 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.015055 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.017503 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.017536 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.017545 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.017563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.017573 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.023664 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" event={"ID":"43d2f41c-af58-4a2a-840a-634408e94c04","Type":"ContainerStarted","Data":"4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.023870 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" event={"ID":"43d2f41c-af58-4a2a-840a-634408e94c04","Type":"ContainerStarted","Data":"3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.023999 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" event={"ID":"43d2f41c-af58-4a2a-840a-634408e94c04","Type":"ContainerStarted","Data":"65f5dd769fd3b003b9b61b326875a2b65691bd57540cd8a44f54e41c318e64bb"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.027200 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/1.log" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.027807 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.032078 5002 scope.go:117] "RemoveContainer" containerID="f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87" Oct 14 07:51:48 crc kubenswrapper[5002]: E1014 07:51:48.032194 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.041721 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.058166 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22f46cac0c8eea38661e0c74156b95a73500536240a17533102e617a22d9876\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:44Z\\\",\\\"message\\\":\\\" 6275 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:51:44.248351 6275 factory.go:656] Stopping watch factory\\\\nI1014 07:51:44.248364 6275 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:51:44.248374 6275 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:51:44.248504 6275 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.248703 6275 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.248881 6275 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249124 6275 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249260 6275 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1014 07:51:44.249522 6275 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:51:44.249585 6275 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.070472 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.081844 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.091652 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.099346 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.112731 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.119663 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.119681 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.119688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.119701 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.119710 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.135680 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.168635 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.188473 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.202579 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.211478 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.221611 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.221668 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.221684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.222034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.222075 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.222807 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.233178 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.242015 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.253911 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.265549 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.281729 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.293145 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.308784 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.320331 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.324112 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.324170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.324188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.324212 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.324230 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.336079 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.347673 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.366913 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.398370 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.416591 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.427376 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.427430 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.427448 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.427467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.427479 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.436262 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.449615 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.530273 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.530607 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.530715 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.530880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.531005 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.568628 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-f5rhq"] Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.569275 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: E1014 07:51:48.569369 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.592052 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.615301 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.630618 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.633919 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.633993 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.634013 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.634040 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.634064 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.652491 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.672315 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.684387 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.684603 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96hv\" (UniqueName: \"kubernetes.io/projected/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-kube-api-access-h96hv\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.735575 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.737294 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.737351 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.737371 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.737401 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.737429 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.758235 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.777163 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.785726 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.785792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96hv\" (UniqueName: \"kubernetes.io/projected/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-kube-api-access-h96hv\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: E1014 07:51:48.786324 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:48 crc kubenswrapper[5002]: E1014 07:51:48.786391 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:49.286369103 +0000 UTC m=+42.267608595 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.801828 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.826235 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96hv\" (UniqueName: \"kubernetes.io/projected/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-kube-api-access-h96hv\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.830927 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.840319 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.840364 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.840384 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.840405 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.840420 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.848243 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.869997 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.886329 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.915360 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.942680 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.943442 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.943509 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.943531 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.943563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.943588 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:48Z","lastTransitionTime":"2025-10-14T07:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.959737 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:48 crc kubenswrapper[5002]: I1014 07:51:48.973708 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.046316 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.046383 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.046411 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.046442 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.046466 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.149341 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.149407 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.149433 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.149464 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.149489 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.252082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.252179 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.252203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.252232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.252253 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.292073 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:49 crc kubenswrapper[5002]: E1014 07:51:49.292252 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:49 crc kubenswrapper[5002]: E1014 07:51:49.292342 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:50.292315526 +0000 UTC m=+43.273555018 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.355861 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.355918 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.355955 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.355984 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.356002 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.459447 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.459537 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.459558 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.459591 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.459616 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.564036 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.564326 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.564460 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.564622 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.564745 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.667949 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.668003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.668019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.668041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.668059 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.719449 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.719521 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:49 crc kubenswrapper[5002]: E1014 07:51:49.719912 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.719578 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:49 crc kubenswrapper[5002]: E1014 07:51:49.720288 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:49 crc kubenswrapper[5002]: E1014 07:51:49.720092 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.770913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.770966 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.770983 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.771006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.771024 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.874244 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.874318 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.874336 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.874358 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.874377 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.977294 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.977345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.977358 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.977377 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:49 crc kubenswrapper[5002]: I1014 07:51:49.977389 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:49Z","lastTransitionTime":"2025-10-14T07:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.080146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.080191 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.080203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.080221 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.080233 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.183202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.183266 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.183284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.183308 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.183331 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.286458 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.286719 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.286815 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.286942 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.287030 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.306287 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:50 crc kubenswrapper[5002]: E1014 07:51:50.306484 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:50 crc kubenswrapper[5002]: E1014 07:51:50.306551 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:52.30653717 +0000 UTC m=+45.287776622 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.389345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.389387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.389399 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.389414 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.389424 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.492546 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.492590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.492601 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.492618 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.492629 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.596495 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.597095 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.597292 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.597469 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.597660 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.702385 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.702796 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.703334 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.703780 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.704178 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.720245 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:50 crc kubenswrapper[5002]: E1014 07:51:50.720438 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.807925 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.808000 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.808022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.808048 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.808067 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.911120 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.911568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.911924 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.912116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:50 crc kubenswrapper[5002]: I1014 07:51:50.912277 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:50Z","lastTransitionTime":"2025-10-14T07:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.015627 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.015691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.015747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.015776 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.015799 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.118113 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.118173 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.118191 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.118217 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.118243 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.220349 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.220409 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.220429 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.220454 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.220471 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.323076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.323142 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.323164 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.323192 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.323215 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.426156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.426204 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.426221 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.426243 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.426298 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.529259 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.529333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.529354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.529386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.529408 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.633217 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.633309 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.633431 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.633506 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.633528 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.720329 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:51 crc kubenswrapper[5002]: E1014 07:51:51.720509 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.720778 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:51 crc kubenswrapper[5002]: E1014 07:51:51.720911 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.721368 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:51 crc kubenswrapper[5002]: E1014 07:51:51.721612 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.736731 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.736865 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.736884 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.736910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.736932 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.839539 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.839625 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.839644 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.839672 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.839690 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.943061 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.943143 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.943166 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.943196 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:51 crc kubenswrapper[5002]: I1014 07:51:51.943219 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:51Z","lastTransitionTime":"2025-10-14T07:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.053595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.053705 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.053727 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.053794 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.053813 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.157592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.157677 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.157690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.157719 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.157735 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.261480 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.261544 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.261561 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.261590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.261608 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.328953 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:52 crc kubenswrapper[5002]: E1014 07:51:52.329228 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:52 crc kubenswrapper[5002]: E1014 07:51:52.329338 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:51:56.329309434 +0000 UTC m=+49.310548916 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.364780 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.364884 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.364900 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.364927 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.364946 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.467766 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.467811 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.467830 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.467899 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.467925 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.570988 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.571067 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.571081 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.571100 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.571113 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.675011 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.675089 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.675109 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.675138 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.675159 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.719971 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:52 crc kubenswrapper[5002]: E1014 07:51:52.720253 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.778770 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.778883 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.778910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.778949 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.778972 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.882902 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.882957 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.882976 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.883001 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.883023 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.985337 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.985372 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.985383 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.985399 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:52 crc kubenswrapper[5002]: I1014 07:51:52.985409 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:52Z","lastTransitionTime":"2025-10-14T07:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.087883 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.087947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.087968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.087999 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.088022 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.191120 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.191192 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.191214 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.191242 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.191264 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.294456 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.294518 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.294534 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.294559 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.294577 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.397606 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.397986 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.398135 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.398274 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.398402 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.502095 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.502474 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.502602 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.502722 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.502882 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.606477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.606529 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.606547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.606573 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.606590 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.709713 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.709755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.709776 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.709798 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.709817 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.720298 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:53 crc kubenswrapper[5002]: E1014 07:51:53.720479 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.720738 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:53 crc kubenswrapper[5002]: E1014 07:51:53.720868 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.721020 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:53 crc kubenswrapper[5002]: E1014 07:51:53.721201 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.813591 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.813975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.814186 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.814333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.814456 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.918445 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.918517 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.918539 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.918573 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:53 crc kubenswrapper[5002]: I1014 07:51:53.918594 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:53Z","lastTransitionTime":"2025-10-14T07:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.021456 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.021519 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.021542 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.021568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.021591 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.032560 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.032802 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.033003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.033177 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.033392 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.051091 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.062810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.063103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.063243 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.063408 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.063629 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.086588 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.092066 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.092161 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.092190 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.092225 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.092248 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.111605 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.116774 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.117781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.118020 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.118172 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.118320 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.137489 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.142174 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.142237 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.142261 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.142288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.142311 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.161537 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.161869 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.164134 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.164192 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.164210 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.164238 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.164255 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.267010 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.267119 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.267140 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.267171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.267189 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.370309 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.370377 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.370394 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.370420 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.370439 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.473690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.473755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.473772 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.473797 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.473814 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.577052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.577125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.577144 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.577161 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.577171 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.679702 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.679764 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.679781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.679804 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.679822 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.719519 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:54 crc kubenswrapper[5002]: E1014 07:51:54.719677 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.783071 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.783170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.783187 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.783206 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.783219 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.886809 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.886878 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.886891 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.886912 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.886926 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.991036 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.991122 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.991146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.991178 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:54 crc kubenswrapper[5002]: I1014 07:51:54.991204 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:54Z","lastTransitionTime":"2025-10-14T07:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.094150 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.094226 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.094250 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.094279 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.094300 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.197676 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.197754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.197768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.197793 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.197805 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.301916 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.302003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.302041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.302077 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.302097 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.405300 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.405344 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.405357 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.405374 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.405383 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.508708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.508771 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.508789 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.508813 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.508830 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.612416 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.612486 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.612505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.612529 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.612547 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.715825 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.715929 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.715947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.715971 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.715989 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.719656 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.719740 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.719664 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:55 crc kubenswrapper[5002]: E1014 07:51:55.719829 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:55 crc kubenswrapper[5002]: E1014 07:51:55.719998 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:55 crc kubenswrapper[5002]: E1014 07:51:55.720152 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.819722 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.819785 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.819818 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.819867 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.819885 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.922979 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.923039 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.923055 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.923078 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:55 crc kubenswrapper[5002]: I1014 07:51:55.923096 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:55Z","lastTransitionTime":"2025-10-14T07:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.026936 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.026998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.027024 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.027053 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.027074 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.129944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.130023 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.130049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.130074 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.130091 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.233670 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.233743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.233762 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.233789 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.233824 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.336646 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.336692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.336708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.336730 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.336748 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.379106 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:56 crc kubenswrapper[5002]: E1014 07:51:56.379298 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:56 crc kubenswrapper[5002]: E1014 07:51:56.379494 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:04.379468948 +0000 UTC m=+57.360708490 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.440440 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.440545 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.440566 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.440591 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.440608 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.544393 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.544452 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.544467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.544492 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.544511 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.648049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.648114 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.648134 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.648158 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.648177 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.719609 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:56 crc kubenswrapper[5002]: E1014 07:51:56.719880 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.751124 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.751184 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.751202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.751226 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.751244 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.855040 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.855110 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.855131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.855229 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.855314 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.958434 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.958494 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.958510 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.958533 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:56 crc kubenswrapper[5002]: I1014 07:51:56.958550 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:56Z","lastTransitionTime":"2025-10-14T07:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.061140 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.061203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.061305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.061334 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.061351 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.164741 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.164878 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.164904 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.164929 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.164948 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.268106 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.268158 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.268175 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.268239 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.268259 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.371280 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.371338 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.371355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.371378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.371395 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.473706 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.473758 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.473777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.473814 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.473830 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.578974 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.579098 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.579123 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.579152 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.579168 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.682507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.682588 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.682615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.682649 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.682672 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.719911 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:57 crc kubenswrapper[5002]: E1014 07:51:57.720080 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.720133 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:57 crc kubenswrapper[5002]: E1014 07:51:57.720308 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.721634 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:57 crc kubenswrapper[5002]: E1014 07:51:57.723119 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.743523 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.764677 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.788970 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.789049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.789071 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.789104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.789127 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.794599 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.812162 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.840432 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.860596 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.882942 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.892267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.892349 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.892367 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.892396 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.892417 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.900125 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.925470 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.945063 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.970539 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.989003 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.996112 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.996150 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.996159 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.996173 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:57 crc kubenswrapper[5002]: I1014 07:51:57.996184 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:57Z","lastTransitionTime":"2025-10-14T07:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.003254 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.014531 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.031467 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.050516 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.063992 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.100333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.100563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.100585 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.101063 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.101271 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.203891 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.203946 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.203956 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.203981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.203992 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.307570 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.307633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.307652 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.307678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.307697 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.410344 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.410382 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.410390 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.410403 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.410412 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.513832 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.513916 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.513934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.513957 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.513978 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.616580 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.616637 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.616654 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.616678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.616696 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.704783 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.709022 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.709153 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.709394 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.709480 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:30.709457319 +0000 UTC m=+83.690696811 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.709673 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:52:30.709640854 +0000 UTC m=+83.690880336 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720565 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720608 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720627 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.720686 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.722204 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.722503 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.728120 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.747574 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.776782 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.794434 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.810532 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.810639 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.810754 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.810802 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.810819 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.810913 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:30.81089215 +0000 UTC m=+83.792131612 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.810758 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.811096 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:30.811067655 +0000 UTC m=+83.792307137 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.811279 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.823755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.823893 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.823914 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.823977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.823999 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.825176 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.847429 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.866226 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.886123 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.903549 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.911548 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.911792 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.911895 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.911924 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:58 crc kubenswrapper[5002]: E1014 07:51:58.912042 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:30.912007932 +0000 UTC m=+83.893247424 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.926829 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.926940 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.926965 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.926994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.927019 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:58Z","lastTransitionTime":"2025-10-14T07:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.937307 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.953333 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.969218 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:58 crc kubenswrapper[5002]: I1014 07:51:58.984171 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:58.999830 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.019238 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:59Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.030089 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.030146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.030166 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.030192 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.030209 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.034296 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:59Z is after 2025-08-24T17:21:41Z" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.133158 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.133323 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.133348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.133383 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.133406 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.236748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.236814 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.236887 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.236923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.236947 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.340033 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.340083 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.340102 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.340129 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.340149 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.443896 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.443944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.443955 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.443972 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.443987 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.546247 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.546298 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.546310 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.546326 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.546339 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.648883 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.648928 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.648939 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.648954 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.648964 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.719726 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.719731 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:51:59 crc kubenswrapper[5002]: E1014 07:51:59.719911 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.719736 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:51:59 crc kubenswrapper[5002]: E1014 07:51:59.720043 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:51:59 crc kubenswrapper[5002]: E1014 07:51:59.720134 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.751700 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.751982 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.752078 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.752166 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.752250 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.855571 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.855637 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.855659 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.855688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.855710 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.958480 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.958525 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.958542 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.958566 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:51:59 crc kubenswrapper[5002]: I1014 07:51:59.958583 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:51:59Z","lastTransitionTime":"2025-10-14T07:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.069053 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.069120 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.069143 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.069171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.069192 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.172616 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.172715 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.172739 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.172769 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.172792 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.275610 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.275664 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.275682 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.275704 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.275723 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.379033 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.379104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.379125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.379157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.379178 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.482227 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.482297 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.482315 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.482339 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.482356 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.586180 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.586244 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.586262 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.586286 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.586304 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.689145 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.689212 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.689229 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.689253 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.689273 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.720090 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:00 crc kubenswrapper[5002]: E1014 07:52:00.720756 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.721183 5002 scope.go:117] "RemoveContainer" containerID="f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.792320 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.792696 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.792707 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.792724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.792735 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.896243 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.896291 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.896307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.896333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.896353 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.998977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.999021 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.999037 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.999058 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:00 crc kubenswrapper[5002]: I1014 07:52:00.999074 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:00Z","lastTransitionTime":"2025-10-14T07:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.089373 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/1.log" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100289 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100858 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100890 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100902 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100917 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100929 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.100997 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.120022 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.143788 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.160143 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.180663 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.202758 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.202791 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.202803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.202820 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.202849 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.207948 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.230673 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.255141 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.268711 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.282087 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.295920 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.305389 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.305423 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.305432 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.305445 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.305455 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.315793 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.328499 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.360617 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.376034 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.393628 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.404546 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.408735 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.408777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.408794 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.408817 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.408834 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.419794 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.441893 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:01Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.511891 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.511956 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.511974 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.511998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.512016 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.615512 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.615582 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.615598 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.615621 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.615641 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.718801 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.718868 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.718880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.718898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.718911 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.719763 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.719816 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:01 crc kubenswrapper[5002]: E1014 07:52:01.719919 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:01 crc kubenswrapper[5002]: E1014 07:52:01.719961 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.719986 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:01 crc kubenswrapper[5002]: E1014 07:52:01.720134 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.821411 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.821487 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.821505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.821531 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.821549 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.924185 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.924258 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.924277 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.924302 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:01 crc kubenswrapper[5002]: I1014 07:52:01.924319 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:01Z","lastTransitionTime":"2025-10-14T07:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.027270 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.027334 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.027355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.027390 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.027415 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.106470 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/2.log" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.107515 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/1.log" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.110500 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" exitCode=1 Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.110560 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.110650 5002 scope.go:117] "RemoveContainer" containerID="f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.111670 5002 scope.go:117] "RemoveContainer" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" Oct 14 07:52:02 crc kubenswrapper[5002]: E1014 07:52:02.112110 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.129825 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.129909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.129929 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.129956 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.129976 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.132967 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.150306 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.169691 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.185130 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.204396 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.225572 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.232782 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.233160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.233352 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.233482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.233612 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.251083 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f79b3b473485d287954af81e545493f10a297d34d8de296f57ae188a7a4d1f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:51:46Z\\\",\\\"message\\\":\\\"te Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-storage-version-migrator-operator/metrics]} name:Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:51:45.936014 6425 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1014 07:51:45.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.266681 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.282759 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.295794 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.311000 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.327112 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.338617 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.338667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.338684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.338708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.338724 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.357743 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.376392 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.399557 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.418716 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.437378 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.441781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.441830 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.441886 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.441918 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.441940 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.454660 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:02Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.544987 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.545044 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.545065 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.545088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.545104 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.648191 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.648232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.648249 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.648269 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.648286 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.719985 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:02 crc kubenswrapper[5002]: E1014 07:52:02.720502 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.751025 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.751082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.751100 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.751121 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.751139 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.854241 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.854280 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.854288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.854303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.854312 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.957345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.957623 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.957696 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.957768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:02 crc kubenswrapper[5002]: I1014 07:52:02.957831 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:02Z","lastTransitionTime":"2025-10-14T07:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.060909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.061149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.061212 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.061271 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.061329 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.116225 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/2.log" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.120134 5002 scope.go:117] "RemoveContainer" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" Oct 14 07:52:03 crc kubenswrapper[5002]: E1014 07:52:03.120359 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.137818 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.149849 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.163503 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.163768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.164088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.164290 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.164472 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.165777 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.176978 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.192921 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.211120 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.228244 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.243313 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.260316 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.267007 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.267062 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.267077 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.267101 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.267113 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.277705 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.303656 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.335077 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.355769 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.370315 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.370350 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.370361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.370378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.370389 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.373085 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.387946 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.406295 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.430905 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.446643 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.473693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.473759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.473776 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.473803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.473823 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.576720 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.576806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.576831 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.576923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.576952 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.679919 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.679994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.680011 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.680038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.680058 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.720044 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.720134 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.720199 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:03 crc kubenswrapper[5002]: E1014 07:52:03.720291 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:03 crc kubenswrapper[5002]: E1014 07:52:03.720410 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:03 crc kubenswrapper[5002]: E1014 07:52:03.720577 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.782823 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.782945 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.782960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.782981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.782997 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.886539 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.886613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.886629 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.886655 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.886671 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.989870 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.989952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.989977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.990006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:03 crc kubenswrapper[5002]: I1014 07:52:03.990025 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:03Z","lastTransitionTime":"2025-10-14T07:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.092950 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.093052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.093073 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.093103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.093122 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.196185 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.196248 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.196267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.196296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.196314 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.299232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.299301 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.299328 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.299356 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.299373 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.402037 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.402123 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.402148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.402176 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.402195 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.425691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.425751 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.425777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.425809 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.425834 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.446032 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:04Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.451152 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.451205 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.451227 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.451254 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.451277 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.470888 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:04Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.475748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.475806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.475827 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.475889 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.475911 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.478812 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.479003 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.479096 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:20.47907214 +0000 UTC m=+73.460311622 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.496517 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:04Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.501056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.501116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.501135 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.501160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.501178 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.522515 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:04Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.528233 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.528286 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.528305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.528331 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.528349 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.547288 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:04Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.547515 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.550322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.550385 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.550409 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.550438 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.550460 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.653941 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.653990 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.654007 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.654031 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.654051 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.720016 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:04 crc kubenswrapper[5002]: E1014 07:52:04.720191 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.757997 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.758074 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.758100 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.758135 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.758156 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.861587 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.861638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.861661 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.861690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.861712 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.964490 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.964569 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.964587 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.964618 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:04 crc kubenswrapper[5002]: I1014 07:52:04.964636 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:04Z","lastTransitionTime":"2025-10-14T07:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.067402 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.067494 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.067523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.067562 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.067585 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.171150 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.171207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.171226 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.171249 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.171266 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.273559 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.273601 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.273611 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.273625 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.273635 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.378158 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.378594 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.378614 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.378638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.378656 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.481908 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.482005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.482028 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.482057 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.482076 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.585291 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.585386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.585412 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.585444 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.585472 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.688602 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.688706 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.688726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.688751 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.688770 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.719459 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.719515 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:05 crc kubenswrapper[5002]: E1014 07:52:05.719715 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.720058 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:05 crc kubenswrapper[5002]: E1014 07:52:05.720252 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:05 crc kubenswrapper[5002]: E1014 07:52:05.720450 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.792167 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.792325 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.792345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.792373 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.792392 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.896386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.896451 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.896468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.896496 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:05 crc kubenswrapper[5002]: I1014 07:52:05.896518 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:05Z","lastTransitionTime":"2025-10-14T07:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.000368 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.000450 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.000483 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.000516 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.000539 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.103982 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.104062 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.104087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.104123 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.104146 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.207215 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.207288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.207312 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.207345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.207369 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.310612 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.310667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.310683 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.310708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.310729 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.414391 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.414450 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.414471 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.414497 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.414517 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.517892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.517972 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.517990 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.518017 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.518034 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.621115 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.621185 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.621211 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.621241 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.621261 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.720364 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:06 crc kubenswrapper[5002]: E1014 07:52:06.720569 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.723704 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.723768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.723791 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.723818 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.723880 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.826809 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.826897 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.826916 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.826940 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.826958 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.929462 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.929530 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.929567 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.929602 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:06 crc kubenswrapper[5002]: I1014 07:52:06.929628 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:06Z","lastTransitionTime":"2025-10-14T07:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.033061 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.033141 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.033171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.033203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.033228 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.142024 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.142087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.142107 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.142131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.142150 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.245005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.245065 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.245083 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.245109 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.245127 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.348188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.348251 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.348268 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.348293 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.348310 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.451735 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.451818 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.451919 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.451988 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.452014 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.555418 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.555465 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.555479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.555497 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.555510 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.658959 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.659029 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.659052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.659166 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.659204 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.720341 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:07 crc kubenswrapper[5002]: E1014 07:52:07.720580 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.721118 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:07 crc kubenswrapper[5002]: E1014 07:52:07.721359 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.721653 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:07 crc kubenswrapper[5002]: E1014 07:52:07.721796 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.759336 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.763668 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.763719 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.763739 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.763763 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.763782 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.777053 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.796619 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.813392 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.836381 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.858980 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.866787 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.866865 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.866880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.866904 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.866921 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.880260 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.897103 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.917078 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.930739 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.960258 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.970157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.970193 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.970203 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.970217 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.970229 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:07Z","lastTransitionTime":"2025-10-14T07:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.976160 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.987340 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:07 crc kubenswrapper[5002]: I1014 07:52:07.996885 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:07Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.010266 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:08Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.022178 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:08Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.042069 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:08Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.055694 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:08Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.073134 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.073177 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.073190 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.073207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.073219 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.175866 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.175912 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.175923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.175939 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.175950 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.278702 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.278781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.278808 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.278874 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.278900 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.383186 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.383263 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.383283 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.383312 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.383331 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.486131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.486209 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.486227 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.486255 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.486275 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.589819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.589915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.589941 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.589978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.590004 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.694635 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.695068 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.695087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.695115 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.695139 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.720475 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:08 crc kubenswrapper[5002]: E1014 07:52:08.720673 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.798091 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.798125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.798139 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.798156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.798170 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.901796 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.901887 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.901907 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.901934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:08 crc kubenswrapper[5002]: I1014 07:52:08.901952 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:08Z","lastTransitionTime":"2025-10-14T07:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.004991 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.005045 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.005062 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.005088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.005107 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.107800 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.107881 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.107900 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.107924 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.107941 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.211132 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.211228 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.211246 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.211633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.211959 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.320507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.320694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.320747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.320781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.320870 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.424902 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.424957 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.424976 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.425000 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.425019 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.528332 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.528394 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.528411 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.528437 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.528459 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.632111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.632182 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.632199 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.632226 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.632245 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.720150 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.720320 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:09 crc kubenswrapper[5002]: E1014 07:52:09.720504 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.720548 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:09 crc kubenswrapper[5002]: E1014 07:52:09.720683 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:09 crc kubenswrapper[5002]: E1014 07:52:09.720875 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.735771 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.735819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.735867 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.735895 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.735915 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.839257 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.839329 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.839351 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.839381 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.839407 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.943191 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.943257 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.943276 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.943302 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:09 crc kubenswrapper[5002]: I1014 07:52:09.943321 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:09Z","lastTransitionTime":"2025-10-14T07:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.045832 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.045918 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.045936 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.045960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.045978 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.148650 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.148726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.148748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.148779 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.148800 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.251908 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.251983 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.252018 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.252040 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.252056 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.355518 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.355571 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.355589 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.355613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.355631 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.459710 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.459786 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.459826 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.459915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.459942 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.563763 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.563930 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.563950 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.563975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.564023 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.667319 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.667382 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.667418 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.667455 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.667479 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.719641 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:10 crc kubenswrapper[5002]: E1014 07:52:10.719925 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.770228 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.770264 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.770278 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.770296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.770308 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.873581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.873613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.873623 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.873638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.873650 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.976109 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.976242 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.976270 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.976304 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:10 crc kubenswrapper[5002]: I1014 07:52:10.976328 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:10Z","lastTransitionTime":"2025-10-14T07:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.079006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.079080 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.079106 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.079135 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.079158 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.181235 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.181273 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.181285 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.181299 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.181309 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.283209 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.283248 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.283260 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.283278 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.283294 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.386264 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.386324 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.386336 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.386385 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.386403 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.489122 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.489273 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.489306 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.489322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.489332 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.592354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.592398 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.592413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.592433 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.592449 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.694602 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.694630 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.694638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.694653 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.694662 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.720549 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.720530 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:11 crc kubenswrapper[5002]: E1014 07:52:11.720756 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.720578 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:11 crc kubenswrapper[5002]: E1014 07:52:11.721024 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:11 crc kubenswrapper[5002]: E1014 07:52:11.721102 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.797500 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.797585 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.797609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.797641 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.797664 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.900728 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.900783 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.900807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.900870 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:11 crc kubenswrapper[5002]: I1014 07:52:11.901048 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:11Z","lastTransitionTime":"2025-10-14T07:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.003810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.003862 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.003871 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.003887 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.003897 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.105893 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.105975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.105994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.106019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.106037 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.208110 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.208160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.208174 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.208195 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.208210 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.310421 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.310479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.310500 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.310524 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.310544 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.412260 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.412296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.412308 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.412329 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.412340 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.514754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.514810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.514824 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.514856 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.514869 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.617613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.617659 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.617671 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.617687 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.617696 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.719791 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:12 crc kubenswrapper[5002]: E1014 07:52:12.720000 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.720424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.720470 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.720479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.720497 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.720506 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.823346 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.823442 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.823461 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.823518 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.823537 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.926331 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.926449 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.926544 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.926607 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:12 crc kubenswrapper[5002]: I1014 07:52:12.926630 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:12Z","lastTransitionTime":"2025-10-14T07:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.029391 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.029482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.029501 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.029589 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.029657 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.134513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.134589 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.134611 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.134642 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.134665 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.237232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.237328 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.237348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.237371 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.237387 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.340106 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.340183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.340207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.340238 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.340262 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.443331 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.443396 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.443413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.443436 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.443451 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.545810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.545871 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.545881 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.545913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.545922 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.648381 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.648468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.648486 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.648512 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.648560 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.720343 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.720402 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.720371 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:13 crc kubenswrapper[5002]: E1014 07:52:13.720581 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:13 crc kubenswrapper[5002]: E1014 07:52:13.720732 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:13 crc kubenswrapper[5002]: E1014 07:52:13.720867 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.750876 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.750910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.750920 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.750934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.750946 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.853877 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.853920 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.853929 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.853945 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.853956 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.956718 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.956764 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.956775 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.956790 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:13 crc kubenswrapper[5002]: I1014 07:52:13.956800 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:13Z","lastTransitionTime":"2025-10-14T07:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.059792 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.059849 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.059860 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.059876 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.059887 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.164737 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.164828 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.164879 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.164907 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.164933 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.267352 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.267392 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.267403 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.267420 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.267450 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.370726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.370762 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.370775 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.370790 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.370802 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.473145 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.473213 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.473232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.473259 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.473278 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.575827 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.575876 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.575887 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.575902 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.575913 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.576770 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.576797 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.576807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.576819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.576827 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.589082 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.592001 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.592047 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.592056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.592073 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.592084 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.603786 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.606307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.606337 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.606346 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.606360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.606369 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.617652 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.621721 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.621747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.621755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.621768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.621776 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.637697 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.641791 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.641898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.641921 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.641991 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.642020 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.656614 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.656757 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.678898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.678931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.678940 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.678958 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.678970 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.719569 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.719679 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.720791 5002 scope.go:117] "RemoveContainer" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" Oct 14 07:52:14 crc kubenswrapper[5002]: E1014 07:52:14.721065 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.781932 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.782114 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.782145 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.782177 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.782201 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.885363 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.885417 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.885434 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.885457 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.885474 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.989193 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.989224 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.989234 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.989250 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:14 crc kubenswrapper[5002]: I1014 07:52:14.989262 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:14Z","lastTransitionTime":"2025-10-14T07:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.090934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.090996 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.091019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.091049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.091071 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.192962 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.193069 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.193087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.193109 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.193125 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.295657 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.295717 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.295735 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.295761 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.295780 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.398262 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.398310 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.398322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.398340 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.398351 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.500960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.501005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.501018 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.501034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.501044 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.603594 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.603695 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.603723 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.603763 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.603803 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.706148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.706191 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.706202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.706219 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.706230 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.719760 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.719806 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:15 crc kubenswrapper[5002]: E1014 07:52:15.719914 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.720005 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:15 crc kubenswrapper[5002]: E1014 07:52:15.720212 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:15 crc kubenswrapper[5002]: E1014 07:52:15.720331 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.809421 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.809488 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.809507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.809535 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.809553 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.912437 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.912527 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.912547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.912574 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:15 crc kubenswrapper[5002]: I1014 07:52:15.912591 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:15Z","lastTransitionTime":"2025-10-14T07:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.015109 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.015171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.015188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.015213 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.015230 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.118817 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.118915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.118934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.118963 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.118981 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.222176 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.222223 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.222235 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.222254 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.222275 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.324985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.325042 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.325060 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.325084 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.325103 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.427346 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.427418 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.427441 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.427477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.427533 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.529183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.529212 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.529222 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.529260 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.529273 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.631508 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.631576 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.631594 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.631621 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.631643 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.719353 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:16 crc kubenswrapper[5002]: E1014 07:52:16.719507 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.733740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.733777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.733787 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.733802 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.733814 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.836376 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.836591 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.836707 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.836798 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.836918 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.939897 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.940079 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.940207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.940359 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:16 crc kubenswrapper[5002]: I1014 07:52:16.940488 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:16Z","lastTransitionTime":"2025-10-14T07:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.043022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.043056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.043064 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.043077 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.043085 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.145087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.145439 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.145604 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.145769 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.145984 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.248952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.249012 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.249028 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.249052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.249069 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.351632 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.351958 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.352058 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.352184 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.352269 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.455623 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.455674 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.455690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.455711 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.455728 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.557874 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.557913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.557925 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.557942 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.557953 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.660337 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.660937 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.660976 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.661006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.661025 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.719657 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:17 crc kubenswrapper[5002]: E1014 07:52:17.719807 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.720105 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:17 crc kubenswrapper[5002]: E1014 07:52:17.720206 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.720452 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:17 crc kubenswrapper[5002]: E1014 07:52:17.720595 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.741727 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.754345 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.764268 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.764330 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.764343 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.764362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.764401 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.767822 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.781654 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.795353 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.811004 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.836754 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.849784 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.862541 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.867058 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.867110 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.867128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.867155 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.867172 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.874485 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.893095 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.911208 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.930561 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.943629 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.969967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.970007 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.970019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.970036 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.970045 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:17Z","lastTransitionTime":"2025-10-14T07:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.973075 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:17 crc kubenswrapper[5002]: I1014 07:52:17.985471 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:17Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.003609 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:18Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.016739 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:18Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.071767 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.071798 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.071808 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.071829 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.071857 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.175272 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.175309 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.175320 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.175337 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.175349 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.278086 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.278128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.278137 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.278154 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.278164 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.381526 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.381572 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.381583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.381600 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.381610 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.489235 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.489296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.489315 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.489346 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.489371 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.592565 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.592623 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.592640 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.592672 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.592690 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.695261 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.695300 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.695313 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.695330 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.695342 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.720298 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:18 crc kubenswrapper[5002]: E1014 07:52:18.720498 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.798036 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.798072 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.798084 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.798103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.798114 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.901450 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.901502 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.901513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.901531 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:18 crc kubenswrapper[5002]: I1014 07:52:18.901544 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:18Z","lastTransitionTime":"2025-10-14T07:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.003795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.003890 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.003911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.003936 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.003988 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.107381 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.107432 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.107449 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.107472 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.107489 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.209523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.209595 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.209612 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.209640 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.209659 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.312341 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.312389 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.312410 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.312437 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.312457 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.414499 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.414593 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.414609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.414633 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.414651 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.517896 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.517951 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.517963 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.517981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.517993 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.620530 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.620562 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.620576 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.620592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.620605 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.720114 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.720288 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:19 crc kubenswrapper[5002]: E1014 07:52:19.720430 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.720730 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:19 crc kubenswrapper[5002]: E1014 07:52:19.720909 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:19 crc kubenswrapper[5002]: E1014 07:52:19.721133 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.724986 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.725355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.725405 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.725440 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.725466 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.828288 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.828319 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.828335 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.828351 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.828362 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.931173 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.931251 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.931275 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.931307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:19 crc kubenswrapper[5002]: I1014 07:52:19.931330 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:19Z","lastTransitionTime":"2025-10-14T07:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.035186 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.035254 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.035270 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.035302 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.035318 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.138971 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.139029 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.139043 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.139064 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.139078 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.242479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.242541 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.242558 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.242584 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.242607 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.345689 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.345750 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.345774 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.345803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.345824 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.449297 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.449380 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.449397 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.449423 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.449441 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.552156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.552253 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.552282 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.552316 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.552336 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.579182 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:20 crc kubenswrapper[5002]: E1014 07:52:20.579400 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:20 crc kubenswrapper[5002]: E1014 07:52:20.579512 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:52:52.579491955 +0000 UTC m=+105.560731407 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.655692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.655753 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.655772 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.655796 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.655813 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.720283 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:20 crc kubenswrapper[5002]: E1014 07:52:20.720466 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.758231 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.758273 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.758285 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.758303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.758315 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.861267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.861317 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.861333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.861357 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.861373 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.964557 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.964599 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.964615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.964639 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:20 crc kubenswrapper[5002]: I1014 07:52:20.964656 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:20Z","lastTransitionTime":"2025-10-14T07:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.066906 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.066955 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.066974 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.066992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.067004 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.169116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.169156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.169168 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.169187 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.169199 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.272653 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.272702 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.272714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.272734 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.272747 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.375742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.375806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.375823 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.375912 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.375943 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.478931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.478992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.479011 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.479035 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.479054 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.582250 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.582316 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.582341 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.582372 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.582394 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.685804 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.685941 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.685967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.685999 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.686025 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.724121 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:21 crc kubenswrapper[5002]: E1014 07:52:21.724300 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.724563 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:21 crc kubenswrapper[5002]: E1014 07:52:21.724654 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.724889 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:21 crc kubenswrapper[5002]: E1014 07:52:21.724978 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.789669 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.789734 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.789759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.789789 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.789812 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.893236 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.893329 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.893355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.893388 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.893412 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.996660 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.996706 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.996724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.996755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:21 crc kubenswrapper[5002]: I1014 07:52:21.996779 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:21Z","lastTransitionTime":"2025-10-14T07:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.100018 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.100096 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.100116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.100149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.100174 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.191104 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/0.log" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.191191 5002 generic.go:334] "Generic (PLEG): container finished" podID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" containerID="4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52" exitCode=1 Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.191240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerDied","Data":"4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.191943 5002 scope.go:117] "RemoveContainer" containerID="4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.205709 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.205770 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.205793 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.205827 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.205887 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.213216 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.237720 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.255643 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.279493 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.304396 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.309628 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.309691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.309708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.309729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.309742 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.323890 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.341975 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.357705 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.374112 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.402526 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.412620 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.412680 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.412694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.412713 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.412725 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.418302 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.431355 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.443635 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.463312 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.481227 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.507068 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.515900 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.515949 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.515960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.515979 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.515992 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.530497 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.545630 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:22Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.618510 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.618547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.618560 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.618581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.618594 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.719450 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:22 crc kubenswrapper[5002]: E1014 07:52:22.719568 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.721066 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.721146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.721172 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.721235 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.721262 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.824308 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.825019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.825073 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.825123 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.825144 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.928350 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.928445 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.928465 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.928490 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:22 crc kubenswrapper[5002]: I1014 07:52:22.928522 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:22Z","lastTransitionTime":"2025-10-14T07:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.031079 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.031122 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.031133 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.031148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.031161 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.134561 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.134615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.134632 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.134655 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.134672 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.198116 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/0.log" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.198189 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerStarted","Data":"35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.216112 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.229489 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.237798 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.238037 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.238166 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.238289 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.238415 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.242561 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.256318 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.269354 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.291492 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.306311 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.333074 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.340973 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.341013 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.341030 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.341053 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.341068 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.348122 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.360946 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.372275 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.383288 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.397453 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.414700 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.429412 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.443689 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.443742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.443751 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.443767 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.443778 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.450037 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.464255 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.497442 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.546338 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.546392 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.546405 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.546424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.546436 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.649590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.649666 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.649684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.649712 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.649730 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.719563 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.719606 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.719626 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:23 crc kubenswrapper[5002]: E1014 07:52:23.719725 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:23 crc kubenswrapper[5002]: E1014 07:52:23.719908 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:23 crc kubenswrapper[5002]: E1014 07:52:23.720067 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.752511 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.752583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.752601 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.752628 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.752650 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.856220 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.856284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.856421 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.856449 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.856821 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.959985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.960046 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.960070 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.960097 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:23 crc kubenswrapper[5002]: I1014 07:52:23.960119 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:23Z","lastTransitionTime":"2025-10-14T07:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.062692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.062742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.062759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.062781 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.062801 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.165232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.165289 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.165311 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.165336 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.165353 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.268874 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.268921 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.268938 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.268957 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.268977 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.372131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.372562 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.372748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.372927 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.373076 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.475938 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.475998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.476015 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.476038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.476056 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.578687 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.578747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.578769 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.578794 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.578813 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.682279 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.682338 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.682355 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.682380 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.682398 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.720166 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:24 crc kubenswrapper[5002]: E1014 07:52:24.720395 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.785037 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.785102 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.785124 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.785154 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.785172 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.888025 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.888089 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.888106 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.888133 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.888152 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.930032 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.930088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.930105 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.930129 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.930145 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: E1014 07:52:24.950156 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:24Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.954808 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.954888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.954906 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.954929 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.954948 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:24 crc kubenswrapper[5002]: E1014 07:52:24.980362 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:24Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.986000 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.986074 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.986099 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.986128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:24 crc kubenswrapper[5002]: I1014 07:52:24.986152 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:24Z","lastTransitionTime":"2025-10-14T07:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.007332 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.018254 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.018338 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.018361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.018391 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.018414 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.040389 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.044951 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.044999 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.045016 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.045043 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.045065 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.066311 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.066537 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.068432 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.068474 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.068489 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.068536 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.068578 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.171867 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.172106 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.172118 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.172139 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.172153 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.275939 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.276013 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.276035 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.276064 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.276087 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.378600 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.378665 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.378690 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.378719 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.378741 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.481969 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.482030 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.482049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.482076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.482097 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.585420 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.585479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.585496 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.585521 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.585541 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.688952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.689021 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.689041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.689071 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.689095 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.719956 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.720000 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.719961 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.720182 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.720316 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:25 crc kubenswrapper[5002]: E1014 07:52:25.720431 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.792381 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.792432 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.792443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.792464 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.792477 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.895678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.895742 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.895766 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.895795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.895821 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.998910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.998986 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.999003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.999023 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:25 crc kubenswrapper[5002]: I1014 07:52:25.999036 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:25Z","lastTransitionTime":"2025-10-14T07:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.101523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.101577 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.101596 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.101622 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.101640 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.205104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.205169 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.205189 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.205216 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.205236 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.308022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.308103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.308131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.308161 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.308181 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.411292 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.411360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.411378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.411405 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.411424 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.514914 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.515024 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.515043 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.515073 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.515092 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.618304 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.618376 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.618404 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.618439 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.618457 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.719488 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:26 crc kubenswrapper[5002]: E1014 07:52:26.719721 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.722052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.722125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.722143 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.722170 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.722188 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.825592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.825649 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.825672 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.825694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.825712 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.929051 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.929107 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.929123 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.929146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:26 crc kubenswrapper[5002]: I1014 07:52:26.929164 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:26Z","lastTransitionTime":"2025-10-14T07:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.032653 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.032710 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.032729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.032754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.032774 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.135329 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.135656 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.135665 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.135696 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.135706 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.238406 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.238480 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.238498 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.238521 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.238538 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.341682 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.341753 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.341771 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.341799 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.341817 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.445427 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.445487 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.445504 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.445527 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.445544 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.548960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.549025 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.549045 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.549073 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.549091 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.652459 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.652526 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.652546 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.652574 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.652594 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.720556 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.720604 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.720617 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:27 crc kubenswrapper[5002]: E1014 07:52:27.720787 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:27 crc kubenswrapper[5002]: E1014 07:52:27.721040 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:27 crc kubenswrapper[5002]: E1014 07:52:27.721184 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.752686 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.756080 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.756131 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.756147 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.756171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.756188 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.773496 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.791286 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.807317 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.827212 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.849740 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.859354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.859419 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.859447 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.859477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.859495 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.866545 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.911951 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.930361 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.948427 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.962581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.962644 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.962663 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.962688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.962707 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:27Z","lastTransitionTime":"2025-10-14T07:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.977505 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:27 crc kubenswrapper[5002]: I1014 07:52:27.995762 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.012599 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.041711 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.057066 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.066297 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.066346 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.066364 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.066388 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.066405 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.082362 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.095642 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.109983 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.169610 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.169675 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.169693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.169717 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.169734 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.272369 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.272422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.272440 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.272463 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.272480 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.376024 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.376096 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.376120 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.376148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.376168 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.479556 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.479625 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.479649 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.479678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.479701 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.582615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.582664 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.582687 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.582714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.582736 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.686305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.686372 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.686392 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.686422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.686448 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.720097 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:28 crc kubenswrapper[5002]: E1014 07:52:28.720359 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.721604 5002 scope.go:117] "RemoveContainer" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.789892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.789958 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.789975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.790001 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.790019 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.892889 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.892964 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.892984 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.893014 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.893033 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.995825 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.995918 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.995939 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.995971 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:28 crc kubenswrapper[5002]: I1014 07:52:28.995994 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:28Z","lastTransitionTime":"2025-10-14T07:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.098880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.098953 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.098978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.099012 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.099034 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.202600 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.202663 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.202710 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.202738 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.202757 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.223018 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/2.log" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.226761 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.227353 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.254548 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.276804 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.304885 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.304923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.304934 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.304952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.304965 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.306422 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.319363 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.336594 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.348419 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.362635 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.375501 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.387888 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.398440 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.407439 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.407477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.407489 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.407507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.407520 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.417508 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.434214 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.447927 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.456935 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.468382 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.481813 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.501868 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.511118 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.511207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.511227 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.511252 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.511271 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.514973 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.614740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.614831 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.614911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.614944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.614967 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.718447 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.718512 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.718532 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.718560 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.718576 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.719828 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.719892 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:29 crc kubenswrapper[5002]: E1014 07:52:29.720046 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.720147 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:29 crc kubenswrapper[5002]: E1014 07:52:29.720320 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:29 crc kubenswrapper[5002]: E1014 07:52:29.720652 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.821676 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.821737 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.821755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.821779 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.821798 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.925163 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.925229 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.925248 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.925275 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:29 crc kubenswrapper[5002]: I1014 07:52:29.925307 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:29Z","lastTransitionTime":"2025-10-14T07:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.028563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.028608 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.028620 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.028638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.028653 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.131909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.131967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.131989 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.132022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.132045 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234060 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/3.log" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234462 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234506 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234518 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234536 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234549 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.234913 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/2.log" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.239516 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" exitCode=1 Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.239574 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.239624 5002 scope.go:117] "RemoveContainer" containerID="7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.240576 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.240866 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.265298 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.285589 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.318949 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7f8c8e02e719087cd885be66413e9bc804aadab39961dd0f353e62108af853fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:01Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI1014 07:52:01.714511 6640 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:52:01.714724 6640 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1014 07:52:01.714739 6640 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1014 07:52:01.714760 6640 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1014 07:52:01.714777 6640 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:52:01.714781 6640 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:52:01.714798 6640 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:52:01.714805 6640 handler.go:208] Removed *v1.Node event handler 2\\\\nI1014 07:52:01.714810 6640 handler.go:208] Removed *v1.Node event handler 7\\\\nI1014 07:52:01.714816 6640 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1014 07:52:01.714822 6640 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:52:01.714933 6640 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:52:01.715021 6640 factory.go:656] Stopping watch factory\\\\nI1014 07:52:01.715052 6640 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:29Z\\\",\\\"message\\\":\\\"console-plugin-85b44fc459-gdk6g openshift-network-operator/iptables-alerter-4ln5h openshift-dns/node-resolver-q57t4 openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-ovn-kubernetes/ovnkube-node-gd5gm]\\\\nI1014 07:52:29.637295 7036 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:52:29.637191 7036 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.338311 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.338360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.338378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.338404 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.338422 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.339443 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.358403 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.379900 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.398638 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.423984 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.441676 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.441895 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.441977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.442004 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.442038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.442064 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.458751 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.499271 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.519468 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.538894 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.545082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.545342 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.545601 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.545767 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.545947 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.555111 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.572338 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.594425 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.613530 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.629695 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.648968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.649021 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.649040 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.649064 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.649085 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.720178 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.720471 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.736967 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.751715 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.751760 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.751777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.751799 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.751819 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.805093 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.805364 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.805448 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.805408269 +0000 UTC m=+147.786647761 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.805586 5002 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.805773 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.805743738 +0000 UTC m=+147.786983260 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.854692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.854729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.854741 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.854759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.854771 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.906122 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.906174 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906312 5002 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906381 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.906364284 +0000 UTC m=+147.887603756 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906475 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906539 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906562 5002 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:52:30 crc kubenswrapper[5002]: E1014 07:52:30.906646 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.906619351 +0000 UTC m=+147.887858833 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.959684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.959745 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.959759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.959778 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:30 crc kubenswrapper[5002]: I1014 07:52:30.959795 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:30Z","lastTransitionTime":"2025-10-14T07:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.007605 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.007816 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.007893 5002 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.007923 5002 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.008033 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.008002516 +0000 UTC m=+147.989241998 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.062897 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.062947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.062958 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.062974 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.062985 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.165360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.165425 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.165443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.165467 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.165484 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.243798 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/3.log" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.246928 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.247168 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.269938 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.270183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.270264 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.270356 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.270456 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.270484 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.289708 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.305169 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.319405 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dba5e20-cacd-4bcb-bc7b-cf2d9517fb41\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8eefae32a67ba6fcd11ac7ad8d4d5a29f3c4197290614149e780404c2fcccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.385760 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.387931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.387977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.387994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.388019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.388036 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.406100 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.425475 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.438160 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.460814 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.473227 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.491489 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.491559 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.491578 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.491605 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.491623 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.492345 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.506467 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.525984 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.544225 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.562776 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.578101 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.595316 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.595361 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.595373 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.595391 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.595404 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.597651 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.613948 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.643724 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:29Z\\\",\\\"message\\\":\\\"console-plugin-85b44fc459-gdk6g openshift-network-operator/iptables-alerter-4ln5h openshift-dns/node-resolver-q57t4 openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-ovn-kubernetes/ovnkube-node-gd5gm]\\\\nI1014 07:52:29.637295 7036 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:52:29.637191 7036 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.697917 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.697967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.697978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.697997 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.698009 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.719564 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.719737 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.719822 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.719941 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.720076 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:31 crc kubenswrapper[5002]: E1014 07:52:31.720408 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.801251 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.801331 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.801354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.801383 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.801400 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.904520 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.904615 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.904638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.904664 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:31 crc kubenswrapper[5002]: I1014 07:52:31.904685 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:31Z","lastTransitionTime":"2025-10-14T07:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.007665 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.008300 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.008507 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.008708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.008894 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.112574 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.112634 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.112656 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.112684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.112704 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.217114 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.217189 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.217206 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.217236 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.217253 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.320797 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.321157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.321362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.321634 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.321922 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.425755 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.425822 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.425882 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.425931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.425955 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.529210 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.529275 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.529295 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.529322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.529342 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.635226 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.635307 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.635327 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.635358 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.635378 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.719678 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:32 crc kubenswrapper[5002]: E1014 07:52:32.719981 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.738460 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.738521 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.738540 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.738565 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.738583 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.841655 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.841718 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.841736 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.841762 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.841779 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.944760 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.944830 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.944875 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.944903 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:32 crc kubenswrapper[5002]: I1014 07:52:32.944921 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:32Z","lastTransitionTime":"2025-10-14T07:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.048139 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.048207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.048225 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.048251 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.048269 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.151754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.151822 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.151888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.151920 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.151938 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.254647 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.254711 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.254728 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.254754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.254773 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.357746 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.357816 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.357833 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.357890 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.357908 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.462268 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.462327 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.462345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.462370 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.462390 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.566167 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.566251 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.566271 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.566301 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.566321 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.669809 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.669921 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.669942 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.669968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.669990 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.721922 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.721977 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:33 crc kubenswrapper[5002]: E1014 07:52:33.722728 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:33 crc kubenswrapper[5002]: E1014 07:52:33.722534 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.722059 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:33 crc kubenswrapper[5002]: E1014 07:52:33.722911 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.773475 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.773534 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.773552 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.773580 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.773600 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.876774 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.876824 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.876873 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.876900 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.876925 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.979825 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.979867 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.979876 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.979891 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:33 crc kubenswrapper[5002]: I1014 07:52:33.979903 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:33Z","lastTransitionTime":"2025-10-14T07:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.082155 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.082215 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.082234 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.082257 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.082275 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.185357 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.185447 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.185463 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.185488 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.185506 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.288581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.288646 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.288664 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.288694 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.288713 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.391967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.392045 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.392063 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.392092 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.392112 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.494869 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.495005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.495025 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.495050 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.495067 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.598118 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.598178 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.598198 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.598225 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.598244 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.701253 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.701701 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.701718 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.701745 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.701771 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.720010 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:34 crc kubenswrapper[5002]: E1014 07:52:34.720251 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.804982 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.805065 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.805088 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.805119 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.805143 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.908736 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.908796 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.908813 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.908863 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:34 crc kubenswrapper[5002]: I1014 07:52:34.908880 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:34Z","lastTransitionTime":"2025-10-14T07:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.012484 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.012546 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.012563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.012590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.012610 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.115613 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.115680 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.115701 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.115729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.115748 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.219111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.219181 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.219198 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.219225 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.219244 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.264305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.264353 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.264369 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.264390 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.264406 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.285400 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.290630 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.290706 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.290748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.290784 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.290812 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.311770 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.317351 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.317733 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.317768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.317800 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.317821 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.338825 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.343923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.343974 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.343995 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.344028 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.344051 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.365964 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.371359 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.371424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.371445 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.371479 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.371512 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.394429 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.394682 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.396710 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.396748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.396763 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.396783 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.396799 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.500054 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.500125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.500149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.500180 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.500201 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.603715 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.603794 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.603832 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.603905 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.603927 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.706295 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.706378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.706395 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.706420 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.706438 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.720018 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.720071 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.720177 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.720263 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.720390 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:35 crc kubenswrapper[5002]: E1014 07:52:35.720653 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.809095 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.809172 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.809194 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.809223 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.809245 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.912373 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.912431 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.912450 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.912473 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:35 crc kubenswrapper[5002]: I1014 07:52:35.912489 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:35Z","lastTransitionTime":"2025-10-14T07:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.015202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.015263 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.015283 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.015306 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.015324 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.119352 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.119426 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.119443 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.119468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.119490 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.221482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.221544 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.221561 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.221587 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.221605 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.324520 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.324617 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.324645 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.324680 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.324705 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.427477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.427533 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.427548 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.427568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.427585 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.530882 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.530972 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.530996 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.531029 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.531052 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.634097 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.634159 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.634177 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.634202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.634222 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.720017 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:36 crc kubenswrapper[5002]: E1014 07:52:36.720532 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.737770 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.737833 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.737909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.737936 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.737956 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.842285 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.842401 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.842422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.842446 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.842463 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.947162 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.947208 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.947224 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.947249 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:36 crc kubenswrapper[5002]: I1014 07:52:36.947268 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:36Z","lastTransitionTime":"2025-10-14T07:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.050451 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.050502 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.050519 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.050542 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.050561 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.153494 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.153549 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.153568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.153594 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.153611 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.256942 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.257006 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.257023 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.257049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.257069 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.359696 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.359733 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.359746 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.359762 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.359773 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.463004 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.463085 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.463116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.463152 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.463173 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.566057 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.566128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.566146 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.566171 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.566187 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.669317 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.669386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.669403 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.669470 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.669491 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.719334 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:37 crc kubenswrapper[5002]: E1014 07:52:37.719468 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.719506 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.720159 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:37 crc kubenswrapper[5002]: E1014 07:52:37.720253 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:37 crc kubenswrapper[5002]: E1014 07:52:37.720421 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.739795 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.754474 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.770427 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.772487 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.772522 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.772532 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.772550 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.772562 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.785023 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.816672 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:29Z\\\",\\\"message\\\":\\\"console-plugin-85b44fc459-gdk6g openshift-network-operator/iptables-alerter-4ln5h openshift-dns/node-resolver-q57t4 openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-ovn-kubernetes/ovnkube-node-gd5gm]\\\\nI1014 07:52:29.637295 7036 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:52:29.637191 7036 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.834608 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.859494 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.870496 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.874236 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.874276 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.874287 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.874306 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.874316 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.885728 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.897269 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.909373 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.918655 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.927219 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.937410 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dba5e20-cacd-4bcb-bc7b-cf2d9517fb41\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8eefae32a67ba6fcd11ac7ad8d4d5a29f3c4197290614149e780404c2fcccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.952830 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.966980 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976558 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976745 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976772 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976780 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.976805 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:37Z","lastTransitionTime":"2025-10-14T07:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:37 crc kubenswrapper[5002]: I1014 07:52:37.997974 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.009776 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:38Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.078813 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.078864 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.078873 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.078888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.078897 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.182360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.182430 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.182449 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.182475 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.182493 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.285598 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.285653 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.285669 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.285693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.285711 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.389436 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.389518 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.389543 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.389577 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.389600 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.493068 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.493132 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.493151 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.493179 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.493198 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.596239 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.596363 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.596446 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.596481 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.596509 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.699975 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.700061 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.700086 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.700118 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.700142 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.719974 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:38 crc kubenswrapper[5002]: E1014 07:52:38.720168 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.803611 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.803691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.803714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.803746 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.803768 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.907750 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.907822 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.907880 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.907909 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:38 crc kubenswrapper[5002]: I1014 07:52:38.907927 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:38Z","lastTransitionTime":"2025-10-14T07:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.010670 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.010724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.010743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.010766 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.010784 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.113619 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.113658 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.113668 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.113683 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.113695 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.216984 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.217041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.217060 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.217090 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.217119 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.321202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.321260 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.321275 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.321295 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.321318 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.425413 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.425511 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.425538 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.425578 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.425618 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.529762 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.529832 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.529889 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.529917 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.529939 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.632952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.633011 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.633027 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.633049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.633065 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.720121 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.720221 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:39 crc kubenswrapper[5002]: E1014 07:52:39.720361 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.720412 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:39 crc kubenswrapper[5002]: E1014 07:52:39.720651 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:39 crc kubenswrapper[5002]: E1014 07:52:39.720726 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.735597 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.735658 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.735675 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.735698 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.735716 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.839183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.839245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.839264 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.839289 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.839309 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.942517 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.942557 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.942568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.942585 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:39 crc kubenswrapper[5002]: I1014 07:52:39.942596 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:39Z","lastTransitionTime":"2025-10-14T07:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.045640 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.045711 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.045749 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.045780 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.045805 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.148056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.148116 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.148130 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.148151 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.148172 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.251678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.251740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.251759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.251783 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.251803 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.355129 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.355193 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.355215 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.355245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.355267 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.458164 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.458232 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.458249 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.458274 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.458293 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.560806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.560903 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.560923 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.560947 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.560964 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.664826 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.664930 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.664944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.664970 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.664988 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.719364 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:40 crc kubenswrapper[5002]: E1014 07:52:40.719591 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.768885 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.768966 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.768994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.769028 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.769049 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.871898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.872004 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.872020 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.872050 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.872073 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.975189 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.975244 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.975256 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.975277 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:40 crc kubenswrapper[5002]: I1014 07:52:40.975290 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:40Z","lastTransitionTime":"2025-10-14T07:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.078485 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.078559 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.078583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.078617 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.078642 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.182452 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.182495 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.182505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.182523 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.182535 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.284795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.284898 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.284918 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.284948 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.284985 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.387711 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.387747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.387777 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.387795 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.387805 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.490924 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.490967 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.490978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.490999 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.491012 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.594583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.594661 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.594688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.594724 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.594750 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.697791 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.697913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.697936 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.697963 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.697981 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.719416 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.719448 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:41 crc kubenswrapper[5002]: E1014 07:52:41.719639 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.719769 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:41 crc kubenswrapper[5002]: E1014 07:52:41.720048 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:41 crc kubenswrapper[5002]: E1014 07:52:41.720131 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.800798 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.800860 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.800873 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.800892 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.800906 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.904093 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.904188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.904214 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.904255 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:41 crc kubenswrapper[5002]: I1014 07:52:41.904277 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:41Z","lastTransitionTime":"2025-10-14T07:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.007034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.007104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.007127 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.007155 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.007251 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.110821 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.110942 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.110960 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.110987 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.111006 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.214333 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.214483 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.214513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.214547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.214572 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.318291 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.318367 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.318387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.318415 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.318433 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.421917 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.422030 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.422092 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.422124 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.422145 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.525922 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.525985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.526008 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.526041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.526064 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.629373 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.629419 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.629430 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.629453 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.629467 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.719412 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:42 crc kubenswrapper[5002]: E1014 07:52:42.719551 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.731531 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.731592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.731609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.731634 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.731651 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.835208 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.835280 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.835293 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.835318 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.835335 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.938868 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.938914 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.938952 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.938973 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:42 crc kubenswrapper[5002]: I1014 07:52:42.938986 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:42Z","lastTransitionTime":"2025-10-14T07:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.042602 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.042667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.042689 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.042721 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.042743 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.145756 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.145807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.145823 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.145884 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.145903 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.248303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.248353 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.248365 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.248385 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.248398 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.351149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.351202 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.351220 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.351245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.351262 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.453908 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.453964 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.453981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.454007 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.454025 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.557303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.557362 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.557375 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.557399 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.557414 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.660591 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.660666 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.660688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.660714 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.660733 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.720080 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.720167 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:43 crc kubenswrapper[5002]: E1014 07:52:43.720286 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.720419 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:43 crc kubenswrapper[5002]: E1014 07:52:43.720442 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:43 crc kubenswrapper[5002]: E1014 07:52:43.720612 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.763968 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.764032 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.764052 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.764079 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.764098 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.867188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.867261 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.867280 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.867305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.867323 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.970388 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.970468 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.970493 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.970526 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:43 crc kubenswrapper[5002]: I1014 07:52:43.970548 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:43Z","lastTransitionTime":"2025-10-14T07:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.072924 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.072987 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.073004 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.073029 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.073047 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.176376 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.176448 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.176465 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.176491 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.176509 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.280046 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.280113 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.280135 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.280161 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.280178 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.383108 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.383178 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.383210 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.383239 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.383260 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.486484 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.486618 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.486643 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.486700 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.486737 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.589873 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.589931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.589944 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.589964 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.589977 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.693421 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.693490 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.693512 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.693538 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.693557 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.719349 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:44 crc kubenswrapper[5002]: E1014 07:52:44.719560 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.797053 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.797107 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.797125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.797149 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.797167 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.918920 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.919004 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.919023 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.919049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:44 crc kubenswrapper[5002]: I1014 07:52:44.919066 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:44Z","lastTransitionTime":"2025-10-14T07:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.022328 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.022454 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.022475 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.022499 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.022516 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.126117 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.126176 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.126194 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.126216 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.126232 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.229921 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.229985 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.230002 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.230026 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.230043 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.333525 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.333635 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.333659 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.333688 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.333709 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.436713 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.436768 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.436785 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.436810 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.436827 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.540282 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.540345 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.540364 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.540389 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.540413 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.635252 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.635683 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.635702 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.635726 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.635743 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.660646 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.666622 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.666684 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.666701 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.666727 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.666744 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.685534 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.690509 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.690568 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.690586 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.690612 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.690630 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.711959 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.717993 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.718108 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.718128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.718183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.718201 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.720395 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.720443 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.720476 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.720599 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.720753 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.720990 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.741344 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.746500 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.746547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.746565 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.746589 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.746607 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.765199 5002 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8e326842-e02f-4877-83ac-d4b3808edbb7\\\",\\\"systemUUID\\\":\\\"a0ba6967-5ef5-4816-9f6e-d1367009bebc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:45 crc kubenswrapper[5002]: E1014 07:52:45.765465 5002 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.771590 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.771626 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.771635 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.771649 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.771659 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.874482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.874545 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.874562 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.874585 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.874603 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.977410 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.977592 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.977609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.977632 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:45 crc kubenswrapper[5002]: I1014 07:52:45.977650 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:45Z","lastTransitionTime":"2025-10-14T07:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.081071 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.081138 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.081157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.081183 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.081202 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.184104 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.184189 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.184207 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.184234 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.184252 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.287315 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.287392 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.287418 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.287449 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.287474 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.390521 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.390579 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.390599 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.390621 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.390637 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.492740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.492800 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.492817 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.492881 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.492901 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.596354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.596424 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.596448 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.596480 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.596503 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.699541 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.699601 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.699618 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.699642 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.699659 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.720465 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:46 crc kubenswrapper[5002]: E1014 07:52:46.720683 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.721772 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 07:52:46 crc kubenswrapper[5002]: E1014 07:52:46.722170 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.802276 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.802348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.802366 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.802396 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.802414 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.904986 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.905049 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.905074 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.905103 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:46 crc kubenswrapper[5002]: I1014 07:52:46.905125 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:46Z","lastTransitionTime":"2025-10-14T07:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.008087 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.008169 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.008194 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.008219 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.008236 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.111911 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.111978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.111997 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.112019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.112036 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.215544 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.215614 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.215631 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.215656 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.215672 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.365572 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.365643 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.365667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.365691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.365707 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.469200 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.469280 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.469318 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.469348 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.469369 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.572547 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.572599 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.572616 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.572641 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.572662 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.675505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.675569 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.675586 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.675610 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.675628 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.720180 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.720246 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.720331 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:47 crc kubenswrapper[5002]: E1014 07:52:47.720515 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:47 crc kubenswrapper[5002]: E1014 07:52:47.721034 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:47 crc kubenswrapper[5002]: E1014 07:52:47.721235 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.742905 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.774930 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"858331e7-9f04-4a5d-8bf9-8307dfa68556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:29Z\\\",\\\"message\\\":\\\"console-plugin-85b44fc459-gdk6g openshift-network-operator/iptables-alerter-4ln5h openshift-dns/node-resolver-q57t4 openshift-kube-apiserver/kube-apiserver-crc openshift-kube-controller-manager/kube-controller-manager-crc openshift-ovn-kubernetes/ovnkube-node-gd5gm]\\\\nI1014 07:52:29.637295 7036 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager-operator/metrics]} name:Service_openshift-kube-controller-manager-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.219:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {3ec9f67e-7758-4707-a6d0-2dc28f28ac37}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:52:29.637191 7036 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96bqb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-gd5gm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.778792 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.778884 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.778906 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.778930 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.778948 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.794753 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.815569 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec8a80e5823d21f54f28d1e87dcf0dc8dc71c92d093a159d05f301ea1d67f0d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.842881 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59c711bea247f698008ad0da0ad4362dba084d86ca621dab617842381ad84783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.868076 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9n4rz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"359daa94-9198-48cf-bbea-a7d8cdb8f3c7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:52:21Z\\\",\\\"message\\\":\\\"2025-10-14T07:51:36+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716\\\\n2025-10-14T07:51:36+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_35501146-0894-40c2-baa8-28dd22c34716 to /host/opt/cni/bin/\\\\n2025-10-14T07:51:36Z [verbose] multus-daemon started\\\\n2025-10-14T07:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wpwqz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9n4rz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.886038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.886981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.887003 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.887027 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.887046 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.888977 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f3de631-7681-4103-8510-2b5545bbe8cd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b08946367938a53a18b0b99237f0152b07b22afb1a28b165c39d4649bda9b73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj84k\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7btf2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.915766 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf44e87d-acc3-427a-b2ce-24daf814d1c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14d28f8908492146c097a50bea489edd563eb84169e3b0d2087b7afc4c64b5ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc220c5f465ba7fc2f9102d8cd0cd36837679a517e93c5261afd2f95e693dbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1de21ae29effda079961a2de10f8b37963bbb8a676a6eba16a735bf13ed39b1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8afa4ae9e2a3d36dbdd128e4b8c9efb37d61397ac032b3553c2961651670a3cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e4ce2b5ee75b4bfeb1f711de3c679a8dae30a94576b76abdbab0c1b3ba6e246\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8f34bf221bb27054b1f8066a1fa0c3dcee3db492d19d1892f46fcc1b82b19c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf9fa340a4590ddfe4b4b058f500823727da14434bc80ccca50dd5ac0889701c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nmkwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-s4jgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.935178 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43d2f41c-af58-4a2a-840a-634408e94c04\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e90025bd686446d61d4dbac9eefa9387bf6c8e91e10b517dd02f836a2f9a92f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ed26bca6b1d64c10bd1cd1dc12d28cb63915f7e1c303293abc0f7b7e8622520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqq27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cpjn4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.951392 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h96hv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:48Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-f5rhq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.968463 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dba5e20-cacd-4bcb-bc7b-cf2d9517fb41\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef8eefae32a67ba6fcd11ac7ad8d4d5a29f3c4197290614149e780404c2fcccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2a08680a00078f6be67af848ba9e4944642d83825bc9eb39e27b59aaad5ea955\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.988378 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f831053-9060-46ae-9a8e-2c6390a15144\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b51674e00f3220ec36df74fc14e6a1e3ca4dbba5032c1e95a7fbb4bda278e69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83093126cf70ab8212c8f87dea729077766f8ec16489e9684ee1544ecb0dfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88a16b3a780a8efc5adbe87731adfd5f12f696ffdbb84133bc7e83be0be04fbd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5fb4c94409c6176fd573414cad672bbfc17ef276fa3722a4f90ff2ccc262389d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:47Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.990177 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.990377 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.990564 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.990715 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:47 crc kubenswrapper[5002]: I1014 07:52:47.990898 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:47Z","lastTransitionTime":"2025-10-14T07:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.009349 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.025920 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lb9sb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"14694d10-47a7-493e-8973-feb8372782ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71f494a8f17bd5464a107774c183d80919367861c905e9be7c069e49ce48344e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sks28\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lb9sb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.055690 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"950a7ddd-ae1b-4183-aead-ac5833e2435f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f458a29ae76fa6a65f9a84b0fba037bf6251d1a5ef2f704d3134ea84592cb1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0a9efa15d8884d8a6b60093b503dc4a3d73539d46899e5b009c79f5f11666f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://829c186de8cd22d6329e0cc2cd356849ce5fdeb69ca734b043e9e98a195d2dec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d71e110fea6408133162d548bd33ab84e121b91ded7572f45468b28d4e2153d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9d862bde8de0dec917b810add71e7718fabaf2c4986340cc58cf86ef7de7ccb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61623a20d52dd130aaf7df2347dcd8ecc23f32f00e24ff61a8d932429de0a630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://239714cc66dd7669ee2c2bd8b160b789eb1ce98253af39161e2998ba3473fbf7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3f47741a4c650a051d74b6fb4240a0d2d12791f052404a4d5307f96c73e1845\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.078244 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db884520-4189-41a9-b630-e1feb6be5ad3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e9ed1f6ebb944a9e33c1b94cc9c436af0b9ff1b8c76c90188df52b82a4cc9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://285fa669d9bc9a5bd198bde957e3da037a0b9807f71089e6cdf8d440de89ae83\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b172438ecb41836e11e1cac107d0fc463862b444c818cf171074d8b431300f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627b7d0740cbf2392d1b8feab0438acca079554fd833d0137314ad2d6a55746\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2374e9dbe50131341f52d86cf3c3561c55a12a0106b659172fd4ec0ba449740a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:51:26Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1014 07:51:21.260376 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:51:21.263119 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2136579338/tls.crt::/tmp/serving-cert-2136579338/tls.key\\\\\\\"\\\\nI1014 07:51:26.950542 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:51:26.955205 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:51:26.955244 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:51:26.955291 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:51:26.955301 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:51:26.966714 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:51:26.966754 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966763 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:51:26.966772 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:51:26.966778 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:51:26.966784 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:51:26.966789 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:51:26.967624 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:51:26.971376 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a898dbf3fe16aebd35df59fc45d6babb08362daf4c03cfc3ed7b1786a7cb06d7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46334df38a6c1f58410680ab9b008b845aa3787ad9b036851fd15758ff6fcb1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.107753 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.107981 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.108136 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.108243 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.108334 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.111389 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138b953d00dcb06409d6ccbdf84cf37af12f401339c72b39d004ae6a368241b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3632f2ebc8853a23f77cf446ee9e85863787cc3eba374ab674a60dd54b2a0850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.153884 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-q57t4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"db1bd93d-cd11-4b20-a5c5-1758b2a53b3c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37059a1372c670426596c2a9403424df5684421e0b1c4559a7eeca5a869368dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2wsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:34Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-q57t4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.169532 5002 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ac70bcb-8cc2-4f05-ac1f-0d1aef5382f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://625ae43a939cac845a2e2c1100c82083c9b90975949f94f76139c34d42aa6035\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9b11ea617c20129cf8f63dc948fac546226fb2b838ad31b2debb4234e4db178\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d59152536ff4486d5933228b94ddec99e7aac0c7604e2c997beb74ce32a59f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://710d485c9dc0823bbc12e218cd648e91ee9d4753172de49d5ff7a108091b6e17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:51:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:51:08Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:51:07Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:52:48Z is after 2025-08-24T17:21:41Z" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.212019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.212339 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.212493 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.212643 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.212800 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.315609 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.315671 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.315693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.315717 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.315735 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.418383 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.418881 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.419111 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.419328 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.419511 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.522305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.522369 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.522389 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.522415 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.522436 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.625636 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.625729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.625754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.625784 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.625806 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.720294 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:48 crc kubenswrapper[5002]: E1014 07:52:48.720724 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.728989 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.729056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.729072 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.729099 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.729116 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.832745 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.832824 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.832881 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.832910 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.832934 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.936732 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.936806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.936827 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.936888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:48 crc kubenswrapper[5002]: I1014 07:52:48.936910 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:48Z","lastTransitionTime":"2025-10-14T07:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.040387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.040460 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.040477 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.040505 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.040523 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.143441 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.143509 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.143528 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.143553 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.143589 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.246913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.247038 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.247110 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.247143 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.247166 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.349740 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.349819 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.349888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.349912 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.349930 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.453195 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.453282 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.453310 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.453342 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.453367 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.557145 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.557228 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.557252 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.557284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.557306 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.660702 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.660788 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.660806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.660834 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.660887 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.720493 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.720620 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:49 crc kubenswrapper[5002]: E1014 07:52:49.720706 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.720488 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:49 crc kubenswrapper[5002]: E1014 07:52:49.720820 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:49 crc kubenswrapper[5002]: E1014 07:52:49.721299 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.764051 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.764360 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.764511 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.764699 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.764867 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.868422 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.868807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.869031 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.869181 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.869330 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.971935 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.972002 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.972019 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.972041 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:49 crc kubenswrapper[5002]: I1014 07:52:49.972057 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:49Z","lastTransitionTime":"2025-10-14T07:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.074883 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.074938 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.074978 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.074999 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.075013 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.178493 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.178553 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.178571 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.178597 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.178614 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.281990 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.282056 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.282078 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.282105 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.282124 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.384612 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.385134 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.385339 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.385757 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.385987 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.491267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.491332 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.491350 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.491377 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.491395 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.594588 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.594644 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.594663 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.594686 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.594702 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.698245 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.698308 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.698327 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.698352 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.698370 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.720042 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:50 crc kubenswrapper[5002]: E1014 07:52:50.720230 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.801583 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.801638 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.801660 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.801692 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.801713 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.905728 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.905802 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.905820 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.905888 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:50 crc kubenswrapper[5002]: I1014 07:52:50.905923 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:50Z","lastTransitionTime":"2025-10-14T07:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.007915 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.007973 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.007992 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.008015 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.008033 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.110241 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.110264 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.110273 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.110284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.110292 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.214082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.214176 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.214194 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.214217 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.214234 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.317482 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.317550 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.317563 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.317587 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.317603 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.420057 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.420156 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.420173 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.420199 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.420223 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.522998 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.523113 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.523140 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.523167 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.523194 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.626515 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.626675 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.626698 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.626721 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.626738 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.719996 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.720099 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.720033 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:51 crc kubenswrapper[5002]: E1014 07:52:51.720233 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:51 crc kubenswrapper[5002]: E1014 07:52:51.720384 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:51 crc kubenswrapper[5002]: E1014 07:52:51.720504 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.728970 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.729034 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.729050 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.729074 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.729091 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.832070 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.832120 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.832137 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.832161 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.832178 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.935402 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.935457 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.935473 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.935498 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:51 crc kubenswrapper[5002]: I1014 07:52:51.935530 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:51Z","lastTransitionTime":"2025-10-14T07:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.039252 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.039330 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.039354 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.039387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.039412 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.147447 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.148214 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.148266 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.148296 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.148317 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.253322 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.253386 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.253403 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.253428 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.253445 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.356210 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.356267 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.356283 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.356305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.356322 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.458594 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.458651 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.458668 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.458691 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.458708 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.562922 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.562977 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.562994 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.563021 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.563038 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.620994 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:52 crc kubenswrapper[5002]: E1014 07:52:52.621213 5002 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:52 crc kubenswrapper[5002]: E1014 07:52:52.621293 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs podName:db613c5e-6bf3-4589-aca8-f6c7da7d36f9 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:56.621269828 +0000 UTC m=+169.602509320 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs") pod "network-metrics-daemon-f5rhq" (UID: "db613c5e-6bf3-4589-aca8-f6c7da7d36f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.666165 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.666261 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.666277 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.666303 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.666322 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.719960 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:52 crc kubenswrapper[5002]: E1014 07:52:52.720198 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.770128 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.770205 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.770221 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.770246 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.770262 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.873673 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.873729 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.873743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.873763 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.873775 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.976543 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.976604 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.976622 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.976713 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:52 crc kubenswrapper[5002]: I1014 07:52:52.976732 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:52Z","lastTransitionTime":"2025-10-14T07:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.080310 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.080368 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.080387 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.080411 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.080429 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.183685 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.183754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.183776 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.183806 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.183831 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.286754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.286829 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.286901 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.286933 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.286958 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.390053 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.390115 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.390133 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.390157 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.390175 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.493803 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.493903 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.493925 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.493950 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.493971 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.596679 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.596733 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.596754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.596783 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.596804 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.699651 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.699722 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.699747 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.699775 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.699798 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.719876 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.719901 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:53 crc kubenswrapper[5002]: E1014 07:52:53.720157 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.720209 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:53 crc kubenswrapper[5002]: E1014 07:52:53.720430 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:53 crc kubenswrapper[5002]: E1014 07:52:53.720576 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.802513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.802581 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.802606 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.802634 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.802655 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.905689 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.905761 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.905782 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.905815 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:53 crc kubenswrapper[5002]: I1014 07:52:53.905864 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:53Z","lastTransitionTime":"2025-10-14T07:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.008436 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.008497 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.008520 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.008553 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.008571 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.111919 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.112005 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.112030 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.112063 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.112091 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.215238 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.215305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.215323 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.215378 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.215408 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.319075 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.319144 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.319160 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.319185 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.319243 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.422678 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.422870 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.422901 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.422931 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.422952 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.526012 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.526076 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.526098 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.526125 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.526144 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.628623 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.628667 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.628686 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.628708 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.628725 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.719669 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:54 crc kubenswrapper[5002]: E1014 07:52:54.720156 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.731454 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.731513 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.731530 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.731552 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.731571 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.834743 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.834807 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.834823 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.834870 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.834887 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.940249 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.940305 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.940324 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.940352 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:54 crc kubenswrapper[5002]: I1014 07:52:54.940372 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:54Z","lastTransitionTime":"2025-10-14T07:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.043371 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.043438 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.043461 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.043494 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.043516 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.145748 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.145811 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.145878 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.145913 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.145937 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.248894 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.248979 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.248997 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.249023 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.249042 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.353185 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.353260 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.353284 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.353316 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.353341 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.455924 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.455995 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.456022 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.456051 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.456073 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.558616 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.558676 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.558693 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.558716 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.558733 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.661754 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.661811 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.661828 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.661886 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.661904 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.720239 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.720310 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.720239 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:55 crc kubenswrapper[5002]: E1014 07:52:55.720440 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:55 crc kubenswrapper[5002]: E1014 07:52:55.720533 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:55 crc kubenswrapper[5002]: E1014 07:52:55.720639 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.765082 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.765148 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.765165 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.765188 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.765206 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.877486 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.877555 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.877580 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.877606 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.877623 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.879485 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.879536 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.879553 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.879574 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.879589 5002 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:52:55Z","lastTransitionTime":"2025-10-14T07:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.951707 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw"] Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.952264 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.958981 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.959050 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.959175 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 14 07:52:55 crc kubenswrapper[5002]: I1014 07:52:55.961300 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.010883 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=90.010811519 podStartE2EDuration="1m30.010811519s" podCreationTimestamp="2025-10-14 07:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.008874217 +0000 UTC m=+108.990113759" watchObservedRunningTime="2025-10-14 07:52:56.010811519 +0000 UTC m=+108.992051011" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.029181 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.029160003 podStartE2EDuration="1m25.029160003s" podCreationTimestamp="2025-10-14 07:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.029011069 +0000 UTC m=+109.010250571" watchObservedRunningTime="2025-10-14 07:52:56.029160003 +0000 UTC m=+109.010399485" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.060174 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.060276 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.060313 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.060389 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.060504 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.072816 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-lb9sb" podStartSLOduration=83.072794594 podStartE2EDuration="1m23.072794594s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.061302801 +0000 UTC m=+109.042542333" watchObservedRunningTime="2025-10-14 07:52:56.072794594 +0000 UTC m=+109.054034056" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.087550 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=58.087525014 podStartE2EDuration="58.087525014s" podCreationTimestamp="2025-10-14 07:51:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.086955578 +0000 UTC m=+109.068195140" watchObservedRunningTime="2025-10-14 07:52:56.087525014 +0000 UTC m=+109.068764506" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.127909 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.127834337 podStartE2EDuration="1m29.127834337s" podCreationTimestamp="2025-10-14 07:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.109311688 +0000 UTC m=+109.090551180" watchObservedRunningTime="2025-10-14 07:52:56.127834337 +0000 UTC m=+109.109073829" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.140707 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-q57t4" podStartSLOduration=83.140682257 podStartE2EDuration="1m23.140682257s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.140110021 +0000 UTC m=+109.121349543" watchObservedRunningTime="2025-10-14 07:52:56.140682257 +0000 UTC m=+109.121921739" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.161420 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.161922 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.162042 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.162177 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.162162 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.162235 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.162270 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.164273 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.167586 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.185128 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66c78ca2-4eac-41c0-a8bf-19ee5315dc8f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7xfgw\" (UID: \"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.217889 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podStartSLOduration=83.217871504 podStartE2EDuration="1m23.217871504s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.2173496 +0000 UTC m=+109.198589082" watchObservedRunningTime="2025-10-14 07:52:56.217871504 +0000 UTC m=+109.199110966" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.243583 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-s4jgg" podStartSLOduration=82.243562322 podStartE2EDuration="1m22.243562322s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.243338506 +0000 UTC m=+109.224577968" watchObservedRunningTime="2025-10-14 07:52:56.243562322 +0000 UTC m=+109.224801784" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.256485 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cpjn4" podStartSLOduration=82.256461132 podStartE2EDuration="1m22.256461132s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.256005271 +0000 UTC m=+109.237244743" watchObservedRunningTime="2025-10-14 07:52:56.256461132 +0000 UTC m=+109.237700624" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.269339 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.279073 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=26.279061279 podStartE2EDuration="26.279061279s" podCreationTimestamp="2025-10-14 07:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.278886784 +0000 UTC m=+109.260126236" watchObservedRunningTime="2025-10-14 07:52:56.279061279 +0000 UTC m=+109.260300741" Oct 14 07:52:56 crc kubenswrapper[5002]: W1014 07:52:56.283208 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66c78ca2_4eac_41c0_a8bf_19ee5315dc8f.slice/crio-e5c112b16fe46afb024f58f8e4b98dc9a7c4bb5d928259a2a225812db187cb81 WatchSource:0}: Error finding container e5c112b16fe46afb024f58f8e4b98dc9a7c4bb5d928259a2a225812db187cb81: Status 404 returned error can't find the container with id e5c112b16fe46afb024f58f8e4b98dc9a7c4bb5d928259a2a225812db187cb81 Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.338222 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9n4rz" podStartSLOduration=82.338202461 podStartE2EDuration="1m22.338202461s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:56.337914332 +0000 UTC m=+109.319153814" watchObservedRunningTime="2025-10-14 07:52:56.338202461 +0000 UTC m=+109.319441913" Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.345075 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" event={"ID":"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f","Type":"ContainerStarted","Data":"e5c112b16fe46afb024f58f8e4b98dc9a7c4bb5d928259a2a225812db187cb81"} Oct 14 07:52:56 crc kubenswrapper[5002]: I1014 07:52:56.720181 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:56 crc kubenswrapper[5002]: E1014 07:52:56.720919 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:57 crc kubenswrapper[5002]: I1014 07:52:57.351680 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" event={"ID":"66c78ca2-4eac-41c0-a8bf-19ee5315dc8f","Type":"ContainerStarted","Data":"09af571046c8428f4a09dd1c92dde7cfd7e61df2d87957854558e27040323df3"} Oct 14 07:52:57 crc kubenswrapper[5002]: I1014 07:52:57.368321 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7xfgw" podStartSLOduration=84.36829011 podStartE2EDuration="1m24.36829011s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:52:57.367555371 +0000 UTC m=+110.348794903" watchObservedRunningTime="2025-10-14 07:52:57.36829011 +0000 UTC m=+110.349529602" Oct 14 07:52:57 crc kubenswrapper[5002]: I1014 07:52:57.720085 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:57 crc kubenswrapper[5002]: I1014 07:52:57.720163 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:57 crc kubenswrapper[5002]: I1014 07:52:57.722622 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:57 crc kubenswrapper[5002]: E1014 07:52:57.722535 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:52:57 crc kubenswrapper[5002]: E1014 07:52:57.723272 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:57 crc kubenswrapper[5002]: E1014 07:52:57.723328 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:58 crc kubenswrapper[5002]: I1014 07:52:58.719742 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:52:58 crc kubenswrapper[5002]: I1014 07:52:58.720810 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 07:52:58 crc kubenswrapper[5002]: E1014 07:52:58.721102 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-gd5gm_openshift-ovn-kubernetes(858331e7-9f04-4a5d-8bf9-8307dfa68556)\"" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" Oct 14 07:52:58 crc kubenswrapper[5002]: E1014 07:52:58.721301 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:52:59 crc kubenswrapper[5002]: I1014 07:52:59.720115 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:52:59 crc kubenswrapper[5002]: I1014 07:52:59.720144 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:52:59 crc kubenswrapper[5002]: E1014 07:52:59.720345 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:52:59 crc kubenswrapper[5002]: E1014 07:52:59.720522 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:52:59 crc kubenswrapper[5002]: I1014 07:52:59.720147 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:52:59 crc kubenswrapper[5002]: E1014 07:52:59.721878 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:00 crc kubenswrapper[5002]: I1014 07:53:00.719350 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:00 crc kubenswrapper[5002]: E1014 07:53:00.719539 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:01 crc kubenswrapper[5002]: I1014 07:53:01.720473 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:01 crc kubenswrapper[5002]: I1014 07:53:01.720576 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:01 crc kubenswrapper[5002]: E1014 07:53:01.720657 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:01 crc kubenswrapper[5002]: E1014 07:53:01.720779 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:01 crc kubenswrapper[5002]: I1014 07:53:01.720938 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:01 crc kubenswrapper[5002]: E1014 07:53:01.721064 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:02 crc kubenswrapper[5002]: I1014 07:53:02.719682 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:02 crc kubenswrapper[5002]: E1014 07:53:02.719943 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:03 crc kubenswrapper[5002]: I1014 07:53:03.719978 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:03 crc kubenswrapper[5002]: I1014 07:53:03.720045 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:03 crc kubenswrapper[5002]: E1014 07:53:03.720153 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:03 crc kubenswrapper[5002]: I1014 07:53:03.720230 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:03 crc kubenswrapper[5002]: E1014 07:53:03.720446 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:03 crc kubenswrapper[5002]: E1014 07:53:03.720786 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:04 crc kubenswrapper[5002]: I1014 07:53:04.720255 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:04 crc kubenswrapper[5002]: E1014 07:53:04.720448 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:05 crc kubenswrapper[5002]: I1014 07:53:05.720306 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:05 crc kubenswrapper[5002]: I1014 07:53:05.720356 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:05 crc kubenswrapper[5002]: E1014 07:53:05.720476 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:05 crc kubenswrapper[5002]: I1014 07:53:05.720556 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:05 crc kubenswrapper[5002]: E1014 07:53:05.720739 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:05 crc kubenswrapper[5002]: E1014 07:53:05.720801 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:06 crc kubenswrapper[5002]: I1014 07:53:06.719478 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:06 crc kubenswrapper[5002]: E1014 07:53:06.719682 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:07 crc kubenswrapper[5002]: I1014 07:53:07.720013 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:07 crc kubenswrapper[5002]: I1014 07:53:07.720067 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:07 crc kubenswrapper[5002]: I1014 07:53:07.720126 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:07 crc kubenswrapper[5002]: E1014 07:53:07.722089 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:07 crc kubenswrapper[5002]: E1014 07:53:07.722471 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:07 crc kubenswrapper[5002]: E1014 07:53:07.722579 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:07 crc kubenswrapper[5002]: E1014 07:53:07.759928 5002 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 14 07:53:07 crc kubenswrapper[5002]: E1014 07:53:07.838758 5002 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.395602 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/1.log" Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.396483 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/0.log" Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.396559 5002 generic.go:334] "Generic (PLEG): container finished" podID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" containerID="35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64" exitCode=1 Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.396604 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerDied","Data":"35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64"} Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.396655 5002 scope.go:117] "RemoveContainer" containerID="4e2a6754b20fbd041d826e3c50216be42d02aec8af48d2aaea8dbf571925ea52" Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.397348 5002 scope.go:117] "RemoveContainer" containerID="35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64" Oct 14 07:53:08 crc kubenswrapper[5002]: E1014 07:53:08.397771 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-9n4rz_openshift-multus(359daa94-9198-48cf-bbea-a7d8cdb8f3c7)\"" pod="openshift-multus/multus-9n4rz" podUID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" Oct 14 07:53:08 crc kubenswrapper[5002]: E1014 07:53:08.419020 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:53:08 crc kubenswrapper[5002]: I1014 07:53:08.719814 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:08 crc kubenswrapper[5002]: E1014 07:53:08.720060 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:09 crc kubenswrapper[5002]: I1014 07:53:09.402531 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/1.log" Oct 14 07:53:09 crc kubenswrapper[5002]: I1014 07:53:09.719949 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:09 crc kubenswrapper[5002]: I1014 07:53:09.720087 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:09 crc kubenswrapper[5002]: I1014 07:53:09.719974 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:09 crc kubenswrapper[5002]: E1014 07:53:09.720210 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:09 crc kubenswrapper[5002]: E1014 07:53:09.720271 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:09 crc kubenswrapper[5002]: E1014 07:53:09.720397 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:10 crc kubenswrapper[5002]: I1014 07:53:10.720243 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:10 crc kubenswrapper[5002]: E1014 07:53:10.720441 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:11 crc kubenswrapper[5002]: I1014 07:53:11.721130 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:11 crc kubenswrapper[5002]: I1014 07:53:11.721200 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:11 crc kubenswrapper[5002]: I1014 07:53:11.721220 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:11 crc kubenswrapper[5002]: E1014 07:53:11.721300 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:11 crc kubenswrapper[5002]: E1014 07:53:11.721399 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:11 crc kubenswrapper[5002]: E1014 07:53:11.721559 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:11 crc kubenswrapper[5002]: I1014 07:53:11.723592 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.415942 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/3.log" Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.419662 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerStarted","Data":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.420311 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.472065 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podStartSLOduration=98.472049969 podStartE2EDuration="1m38.472049969s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:12.469039209 +0000 UTC m=+125.450278691" watchObservedRunningTime="2025-10-14 07:53:12.472049969 +0000 UTC m=+125.453289431" Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.719775 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:12 crc kubenswrapper[5002]: E1014 07:53:12.719949 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:12 crc kubenswrapper[5002]: I1014 07:53:12.784710 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-f5rhq"] Oct 14 07:53:12 crc kubenswrapper[5002]: E1014 07:53:12.840129 5002 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:53:13 crc kubenswrapper[5002]: I1014 07:53:13.423096 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:13 crc kubenswrapper[5002]: E1014 07:53:13.423229 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:13 crc kubenswrapper[5002]: I1014 07:53:13.720002 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:13 crc kubenswrapper[5002]: I1014 07:53:13.720176 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:13 crc kubenswrapper[5002]: I1014 07:53:13.720370 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:13 crc kubenswrapper[5002]: E1014 07:53:13.720746 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:13 crc kubenswrapper[5002]: E1014 07:53:13.720952 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:13 crc kubenswrapper[5002]: E1014 07:53:13.721137 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:14 crc kubenswrapper[5002]: I1014 07:53:14.720250 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:14 crc kubenswrapper[5002]: E1014 07:53:14.720466 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:15 crc kubenswrapper[5002]: I1014 07:53:15.720321 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:15 crc kubenswrapper[5002]: I1014 07:53:15.720385 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:15 crc kubenswrapper[5002]: I1014 07:53:15.720330 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:15 crc kubenswrapper[5002]: E1014 07:53:15.720573 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:15 crc kubenswrapper[5002]: E1014 07:53:15.720698 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:15 crc kubenswrapper[5002]: E1014 07:53:15.720737 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:16 crc kubenswrapper[5002]: I1014 07:53:16.720251 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:16 crc kubenswrapper[5002]: E1014 07:53:16.720439 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:17 crc kubenswrapper[5002]: I1014 07:53:17.136536 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 07:53:17 crc kubenswrapper[5002]: I1014 07:53:17.719501 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:17 crc kubenswrapper[5002]: I1014 07:53:17.719564 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:17 crc kubenswrapper[5002]: I1014 07:53:17.719576 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:17 crc kubenswrapper[5002]: E1014 07:53:17.721764 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:17 crc kubenswrapper[5002]: E1014 07:53:17.721914 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:17 crc kubenswrapper[5002]: E1014 07:53:17.722047 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:17 crc kubenswrapper[5002]: E1014 07:53:17.840927 5002 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:53:18 crc kubenswrapper[5002]: E1014 07:53:18.499989 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:53:18 crc kubenswrapper[5002]: I1014 07:53:18.719829 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:18 crc kubenswrapper[5002]: E1014 07:53:18.720098 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:19 crc kubenswrapper[5002]: I1014 07:53:19.720097 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:19 crc kubenswrapper[5002]: E1014 07:53:19.720673 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:19 crc kubenswrapper[5002]: I1014 07:53:19.720147 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:19 crc kubenswrapper[5002]: E1014 07:53:19.720956 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:19 crc kubenswrapper[5002]: I1014 07:53:19.720098 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:19 crc kubenswrapper[5002]: E1014 07:53:19.721192 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:20 crc kubenswrapper[5002]: I1014 07:53:20.720028 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:20 crc kubenswrapper[5002]: I1014 07:53:20.720976 5002 scope.go:117] "RemoveContainer" containerID="35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64" Oct 14 07:53:20 crc kubenswrapper[5002]: E1014 07:53:20.723061 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:21 crc kubenswrapper[5002]: I1014 07:53:21.455658 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/1.log" Oct 14 07:53:21 crc kubenswrapper[5002]: I1014 07:53:21.456135 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerStarted","Data":"936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a"} Oct 14 07:53:21 crc kubenswrapper[5002]: I1014 07:53:21.720309 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:21 crc kubenswrapper[5002]: E1014 07:53:21.720497 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:53:21 crc kubenswrapper[5002]: I1014 07:53:21.720536 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:21 crc kubenswrapper[5002]: E1014 07:53:21.720686 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:53:21 crc kubenswrapper[5002]: I1014 07:53:21.721042 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:21 crc kubenswrapper[5002]: E1014 07:53:21.721260 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:53:22 crc kubenswrapper[5002]: I1014 07:53:22.719790 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:22 crc kubenswrapper[5002]: E1014 07:53:22.720277 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f5rhq" podUID="db613c5e-6bf3-4589-aca8-f6c7da7d36f9" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.720270 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.720307 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.720477 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.723682 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.723935 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.725102 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 14 07:53:23 crc kubenswrapper[5002]: I1014 07:53:23.725599 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 14 07:53:24 crc kubenswrapper[5002]: I1014 07:53:24.720048 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:24 crc kubenswrapper[5002]: I1014 07:53:24.722609 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 14 07:53:24 crc kubenswrapper[5002]: I1014 07:53:24.723004 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.862759 5002 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.916340 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd9r6"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.917446 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.933746 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.933873 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.934666 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.936693 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.936693 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.936985 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.937079 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.937528 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.937751 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.938425 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.940621 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.941589 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.943158 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ml5x5"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.946310 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.949958 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.950084 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.950874 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.951037 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.951318 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.951724 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.952600 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.955923 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.958238 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.959473 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.961156 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.979456 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dg9pb"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.980006 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hjt8h"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.980883 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981157 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981198 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981456 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981641 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981756 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.981897 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.982639 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.982814 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.983080 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.983225 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.984107 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.985093 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg"] Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.985192 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.985626 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.992273 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.992598 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.992738 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.992765 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.999418 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 14 07:53:26 crc kubenswrapper[5002]: I1014 07:53:26.999944 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000120 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000276 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000343 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000587 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000629 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000773 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000953 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001045 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001110 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001289 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001478 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001499 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001568 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001734 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.001898 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.002474 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.002953 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003153 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003190 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003277 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003301 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003402 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.000774 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003527 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003648 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003747 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003869 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.003986 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.004083 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.004195 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.004292 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.004408 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.004996 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.007314 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.011274 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.011894 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.017064 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.017457 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q9jk6"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.017652 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.017898 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.017933 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.018234 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.018406 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.019644 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.019954 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.020298 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.020320 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021163 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021305 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-client\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021330 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021346 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8cbb\" (UniqueName: \"kubernetes.io/projected/6b840956-6900-4c7c-89ce-622ddf455cf6-kube-api-access-g8cbb\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021361 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-service-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021376 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021389 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021410 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021426 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021441 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021456 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-machine-approver-tls\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021476 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-audit\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021490 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021502 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-trusted-ca\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021515 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-encryption-config\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021596 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-serving-cert\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021632 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021655 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021683 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/841fda0e-a272-490f-81ab-dc83e48b2201-serving-cert\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021708 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021730 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433fc6e4-cadb-45b7-964b-cd0a82c09897-serving-cert\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021751 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021779 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-node-pullsecrets\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021799 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021820 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021858 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npp2w\" (UniqueName: \"kubernetes.io/projected/433fc6e4-cadb-45b7-964b-cd0a82c09897-kube-api-access-npp2w\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021890 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-encryption-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021920 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qw86\" (UniqueName: \"kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.021941 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022065 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thn47\" (UniqueName: \"kubernetes.io/projected/3fdd526f-086e-4fb1-8a38-d297ae1843e6-kube-api-access-thn47\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022086 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022110 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmgs\" (UniqueName: \"kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022124 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022146 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022159 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-auth-proxy-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022174 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-client\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022188 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022202 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-images\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022214 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022227 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9bcc50b-a9bb-4ccf-bb92-42718291689d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022245 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-client\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022258 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-config\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022272 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-serving-cert\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022287 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gnm\" (UniqueName: \"kubernetes.io/projected/42350c04-b619-45e9-8567-8f71d4c95e46-kube-api-access-n9gnm\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022310 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022325 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022349 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022363 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-config\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022377 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msvb4\" (UniqueName: \"kubernetes.io/projected/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-kube-api-access-msvb4\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022393 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022407 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-config\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022439 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-image-import-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022453 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ndd5\" (UniqueName: \"kubernetes.io/projected/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-kube-api-access-9ndd5\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022492 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022524 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lqn9\" (UniqueName: \"kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022541 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-service-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022556 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-policies\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022571 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b840956-6900-4c7c-89ce-622ddf455cf6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022585 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022603 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jsrt\" (UniqueName: \"kubernetes.io/projected/841fda0e-a272-490f-81ab-dc83e48b2201-kube-api-access-4jsrt\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022629 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022660 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4m5d\" (UniqueName: \"kubernetes.io/projected/7ea0cd04-9117-4160-bb93-e211cf1cba70-kube-api-access-b4m5d\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022674 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022689 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022734 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-dir\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022806 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-srv-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022902 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022932 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bcc50b-a9bb-4ccf-bb92-42718291689d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022954 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qc94\" (UniqueName: \"kubernetes.io/projected/f9bcc50b-a9bb-4ccf-bb92-42718291689d-kube-api-access-9qc94\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.022998 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023022 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx4wp\" (UniqueName: \"kubernetes.io/projected/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-kube-api-access-wx4wp\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023062 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023086 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-audit-dir\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023112 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023154 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkbmz\" (UniqueName: \"kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023176 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-config\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023205 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023227 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt7rl\" (UniqueName: \"kubernetes.io/projected/01c76e0a-94db-4229-8201-dfb777a5ccd0-kube-api-access-qt7rl\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.023294 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.025418 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-serving-cert\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.025523 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.027762 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.029518 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.030133 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.033504 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.033798 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.034106 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.034586 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.035443 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.036462 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.036718 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.036805 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.036911 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.037017 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.041044 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.041186 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.051177 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.054626 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.059349 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fv4sf"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.059830 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.060085 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s656d"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.060240 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.060490 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.060613 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.060941 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.061010 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.061779 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.062430 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.062600 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.062707 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.062808 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.062978 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063099 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063215 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063337 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063700 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063756 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.063916 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.064139 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.064344 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.064483 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.064591 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.064666 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjgv5"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.065115 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.065267 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.065500 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.065763 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.066672 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.067137 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.067289 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.067723 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.068019 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.068212 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.066630 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.073300 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.073884 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.079684 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.083722 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.084289 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-l9m77"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.087021 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.087252 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.092007 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.092518 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.093887 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.094222 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.094850 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.095374 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.099374 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.100997 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.103077 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.110632 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wmg78"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.113190 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.114266 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.114662 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.114808 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.120189 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-dndgk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.121299 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.123201 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.124091 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126317 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jsrt\" (UniqueName: \"kubernetes.io/projected/841fda0e-a272-490f-81ab-dc83e48b2201-kube-api-access-4jsrt\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126348 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-config\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126463 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126502 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126528 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzss2\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-kube-api-access-nzss2\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126563 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126582 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bcc50b-a9bb-4ccf-bb92-42718291689d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126600 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qc94\" (UniqueName: \"kubernetes.io/projected/f9bcc50b-a9bb-4ccf-bb92-42718291689d-kube-api-access-9qc94\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126616 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-audit-dir\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126671 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-audit-dir\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126765 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.126947 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.127996 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bcc50b-a9bb-4ccf-bb92-42718291689d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128010 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128438 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128471 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128494 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/99337d25-bf3b-4740-9e30-8ef50cecd6cd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128534 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt7rl\" (UniqueName: \"kubernetes.io/projected/01c76e0a-94db-4229-8201-dfb777a5ccd0-kube-api-access-qt7rl\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128551 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-config\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128569 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128586 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-client\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128609 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128630 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-service-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128659 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128681 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128707 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128733 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128758 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128781 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128804 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128825 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-trusted-ca\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128866 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-encryption-config\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128883 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20a25551-3a54-4eb2-8e5a-e13134f2ea95-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128901 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-audit\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128925 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128955 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v9dw\" (UniqueName: \"kubernetes.io/projected/5b600621-f833-48ef-9d9d-04f6543d81ae-kube-api-access-4v9dw\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128975 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.128978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129010 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-serving-cert\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129029 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-profile-collector-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129047 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npp2w\" (UniqueName: \"kubernetes.io/projected/433fc6e4-cadb-45b7-964b-cd0a82c09897-kube-api-access-npp2w\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129062 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129079 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qw86\" (UniqueName: \"kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129094 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129114 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qgb2\" (UniqueName: \"kubernetes.io/projected/9aa1836d-6217-4c0f-bb26-dfc40643305d-kube-api-access-2qgb2\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-auth-proxy-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129165 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de309abf-bc67-416d-b0b0-1da8f0e952a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129173 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xdlrk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129188 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-webhook-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129214 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129239 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-images\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129268 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26k5p\" (UniqueName: \"kubernetes.io/projected/de309abf-bc67-416d-b0b0-1da8f0e952a2-kube-api-access-26k5p\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129292 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129331 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-config\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129364 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129388 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2rjk\" (UniqueName: \"kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129405 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4smz\" (UniqueName: \"kubernetes.io/projected/77e57d6f-a931-4fee-a933-2abb2cbe22d3-kube-api-access-s4smz\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-serving-cert\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129439 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b689e3e-c89a-42bc-b6df-180d587cc197-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129457 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gnm\" (UniqueName: \"kubernetes.io/projected/42350c04-b619-45e9-8567-8f71d4c95e46-kube-api-access-n9gnm\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129472 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129680 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/007477a4-2f7b-46a0-9254-fe1da6f900f4-tmpfs\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129697 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a25551-3a54-4eb2-8e5a-e13134f2ea95-serving-cert\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129717 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msvb4\" (UniqueName: \"kubernetes.io/projected/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-kube-api-access-msvb4\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129733 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129748 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-config\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129764 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b600621-f833-48ef-9d9d-04f6543d81ae-proxy-tls\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129782 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129799 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ndd5\" (UniqueName: \"kubernetes.io/projected/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-kube-api-access-9ndd5\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129818 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4726\" (UniqueName: \"kubernetes.io/projected/ce76625d-d545-4664-b365-97cd648cb7ab-kube-api-access-l4726\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129849 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-service-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129866 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129886 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b840956-6900-4c7c-89ce-622ddf455cf6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129906 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129927 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129942 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129962 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-dir\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129983 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-srv-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130001 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4m5d\" (UniqueName: \"kubernetes.io/projected/7ea0cd04-9117-4160-bb93-e211cf1cba70-kube-api-access-b4m5d\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130018 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130036 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130052 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx4wp\" (UniqueName: \"kubernetes.io/projected/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-kube-api-access-wx4wp\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130069 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slhr5\" (UniqueName: \"kubernetes.io/projected/99337d25-bf3b-4740-9e30-8ef50cecd6cd-kube-api-access-slhr5\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130085 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b689e3e-c89a-42bc-b6df-180d587cc197-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130100 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkbmz\" (UniqueName: \"kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130116 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130133 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130150 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130167 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-serving-cert\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130182 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-srv-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130197 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/059e5978-1a83-45f0-876a-c497fa2365b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130214 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8cbb\" (UniqueName: \"kubernetes.io/projected/6b840956-6900-4c7c-89ce-622ddf455cf6-kube-api-access-g8cbb\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130229 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de309abf-bc67-416d-b0b0-1da8f0e952a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130247 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfzxt\" (UniqueName: \"kubernetes.io/projected/20a25551-3a54-4eb2-8e5a-e13134f2ea95-kube-api-access-qfzxt\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130264 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v79pj\" (UniqueName: \"kubernetes.io/projected/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-kube-api-access-v79pj\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130281 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130298 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130316 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-machine-approver-tls\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130333 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgt7z\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-kube-api-access-dgt7z\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130351 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wbq2\" (UniqueName: \"kubernetes.io/projected/89e82ba1-52a9-4c22-bc43-7d71bf835d2f-kube-api-access-8wbq2\") pod \"migrator-59844c95c7-4r89t\" (UID: \"89e82ba1-52a9-4c22-bc43-7d71bf835d2f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130376 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130393 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/841fda0e-a272-490f-81ab-dc83e48b2201-serving-cert\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130409 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppcn8\" (UniqueName: \"kubernetes.io/projected/007477a4-2f7b-46a0-9254-fe1da6f900f4-kube-api-access-ppcn8\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130439 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-key\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130458 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130473 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130490 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433fc6e4-cadb-45b7-964b-cd0a82c09897-serving-cert\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130522 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-config\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130538 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-node-pullsecrets\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130555 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130580 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-encryption-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130596 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thn47\" (UniqueName: \"kubernetes.io/projected/3fdd526f-086e-4fb1-8a38-d297ae1843e6-kube-api-access-thn47\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130612 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmgs\" (UniqueName: \"kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130647 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/059e5978-1a83-45f0-876a-c497fa2365b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130664 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130680 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ce76625d-d545-4664-b365-97cd648cb7ab-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130697 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4hj5\" (UniqueName: \"kubernetes.io/projected/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-kube-api-access-q4hj5\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130713 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130730 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-client\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130746 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b600621-f833-48ef-9d9d-04f6543d81ae-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130761 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-cabundle\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130778 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-client\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130794 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9bcc50b-a9bb-4ccf-bb92-42718291689d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130810 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.130827 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-apiservice-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.131297 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.131302 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-service-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.131316 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.131362 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-serving-cert\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.132052 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-auth-proxy-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.132449 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-trusted-ca\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.132582 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.133238 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.133622 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.133999 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.134148 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.135092 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.135181 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/42350c04-b619-45e9-8567-8f71d4c95e46-node-pullsecrets\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.135213 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.135282 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.135307 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.136469 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.137248 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-images\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.137369 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-serving-cert\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.137679 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-encryption-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.137705 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-config\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.137997 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.138226 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-ca\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.138278 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.138312 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.139150 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.139441 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-config\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.139578 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7ea0cd04-9117-4160-bb93-e211cf1cba70-srv-cert\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140149 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140307 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd9r6"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140412 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140474 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-serving-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140659 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.129540 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.140414 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-service-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.141392 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.141742 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-config\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.142660 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-dir\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.142809 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-audit\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.143054 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.143979 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.144410 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-config\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.144790 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.144948 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/433fc6e4-cadb-45b7-964b-cd0a82c09897-config\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.145472 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-serving-cert\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.145686 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.145827 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-encryption-config\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.145879 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146066 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-client\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146068 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-machine-approver-tls\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146115 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146255 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-l8k2s"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146282 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9aa1836d-6217-4c0f-bb26-dfc40643305d-metrics-tls\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146312 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146337 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146356 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-config\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146368 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146375 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146624 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-image-import-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146655 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4nn\" (UniqueName: \"kubernetes.io/projected/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-kube-api-access-jl4nn\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146678 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.146827 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147014 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lqn9\" (UniqueName: \"kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147036 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzjvs\" (UniqueName: \"kubernetes.io/projected/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-kube-api-access-fzjvs\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147058 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-policies\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147085 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147403 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dg9pb"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147519 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.147595 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9bcc50b-a9bb-4ccf-bb92-42718291689d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.148696 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149166 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-image-import-ca\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149234 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149253 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/841fda0e-a272-490f-81ab-dc83e48b2201-serving-cert\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149266 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149413 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6b840956-6900-4c7c-89ce-622ddf455cf6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149468 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/433fc6e4-cadb-45b7-964b-cd0a82c09897-serving-cert\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149706 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/841fda0e-a272-490f-81ab-dc83e48b2201-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.149968 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.150124 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hjt8h"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.150481 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3fdd526f-086e-4fb1-8a38-d297ae1843e6-audit-policies\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.151057 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjgv5"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.151382 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-config\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.152455 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.152706 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-etcd-client\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.154054 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.157799 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.158280 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.158410 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.160540 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42350c04-b619-45e9-8567-8f71d4c95e46-serving-cert\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.163764 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.164738 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/42350c04-b619-45e9-8567-8f71d4c95e46-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.166538 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s656d"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.168063 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.170419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/01c76e0a-94db-4229-8201-dfb777a5ccd0-etcd-client\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.175862 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.175950 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wmg78"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.177240 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.180130 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.181481 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-s6s5j"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.182027 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.182071 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.182605 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cnwsm"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.183594 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.183607 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.185538 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q9jk6"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.186512 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.187472 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.188484 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.189691 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.191088 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fv4sf"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.192926 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ml5x5"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.194317 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.197330 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.199932 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.200182 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.200995 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.204050 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.206003 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.207665 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.211881 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.211907 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.213627 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cnwsm"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.214708 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.216681 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-l9m77"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.219322 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.221633 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.221751 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.228920 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.228956 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s6s5j"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.231873 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xdlrk"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.232537 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rjkc7"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.233344 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.235516 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rjkc7"] Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.239906 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247634 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-profile-collector-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247678 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qgb2\" (UniqueName: \"kubernetes.io/projected/9aa1836d-6217-4c0f-bb26-dfc40643305d-kube-api-access-2qgb2\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247697 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de309abf-bc67-416d-b0b0-1da8f0e952a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247714 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-webhook-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247730 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26k5p\" (UniqueName: \"kubernetes.io/projected/de309abf-bc67-416d-b0b0-1da8f0e952a2-kube-api-access-26k5p\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247747 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247765 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2rjk\" (UniqueName: \"kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247780 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4smz\" (UniqueName: \"kubernetes.io/projected/77e57d6f-a931-4fee-a933-2abb2cbe22d3-kube-api-access-s4smz\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247796 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b689e3e-c89a-42bc-b6df-180d587cc197-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247817 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a25551-3a54-4eb2-8e5a-e13134f2ea95-serving-cert\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247849 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/007477a4-2f7b-46a0-9254-fe1da6f900f4-tmpfs\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247873 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b600621-f833-48ef-9d9d-04f6543d81ae-proxy-tls\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247897 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4726\" (UniqueName: \"kubernetes.io/projected/ce76625d-d545-4664-b365-97cd648cb7ab-kube-api-access-l4726\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247915 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247932 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247958 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slhr5\" (UniqueName: \"kubernetes.io/projected/99337d25-bf3b-4740-9e30-8ef50cecd6cd-kube-api-access-slhr5\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247977 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b689e3e-c89a-42bc-b6df-180d587cc197-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.247992 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248006 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-srv-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248020 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/059e5978-1a83-45f0-876a-c497fa2365b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248039 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de309abf-bc67-416d-b0b0-1da8f0e952a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248053 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfzxt\" (UniqueName: \"kubernetes.io/projected/20a25551-3a54-4eb2-8e5a-e13134f2ea95-kube-api-access-qfzxt\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248070 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v79pj\" (UniqueName: \"kubernetes.io/projected/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-kube-api-access-v79pj\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248085 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgt7z\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-kube-api-access-dgt7z\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248111 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wbq2\" (UniqueName: \"kubernetes.io/projected/89e82ba1-52a9-4c22-bc43-7d71bf835d2f-kube-api-access-8wbq2\") pod \"migrator-59844c95c7-4r89t\" (UID: \"89e82ba1-52a9-4c22-bc43-7d71bf835d2f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248126 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppcn8\" (UniqueName: \"kubernetes.io/projected/007477a4-2f7b-46a0-9254-fe1da6f900f4-kube-api-access-ppcn8\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248157 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-key\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248171 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248186 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-config\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248218 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/059e5978-1a83-45f0-876a-c497fa2365b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248234 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ce76625d-d545-4664-b365-97cd648cb7ab-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248249 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4hj5\" (UniqueName: \"kubernetes.io/projected/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-kube-api-access-q4hj5\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248265 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b600621-f833-48ef-9d9d-04f6543d81ae-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248280 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-cabundle\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248295 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248335 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248357 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-apiservice-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248373 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248390 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-serving-cert\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248408 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9aa1836d-6217-4c0f-bb26-dfc40643305d-metrics-tls\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248426 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4nn\" (UniqueName: \"kubernetes.io/projected/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-kube-api-access-jl4nn\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248449 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzjvs\" (UniqueName: \"kubernetes.io/projected/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-kube-api-access-fzjvs\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248476 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-config\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248493 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzss2\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-kube-api-access-nzss2\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248528 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/99337d25-bf3b-4740-9e30-8ef50cecd6cd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248553 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248569 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248584 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248600 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20a25551-3a54-4eb2-8e5a-e13134f2ea95-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248624 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v9dw\" (UniqueName: \"kubernetes.io/projected/5b600621-f833-48ef-9d9d-04f6543d81ae-kube-api-access-4v9dw\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.248638 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.249808 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/007477a4-2f7b-46a0-9254-fe1da6f900f4-tmpfs\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.250166 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/20a25551-3a54-4eb2-8e5a-e13134f2ea95-available-featuregates\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.250465 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5b600621-f833-48ef-9d9d-04f6543d81ae-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.250605 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.250666 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.252426 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-profile-collector-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.259264 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.267627 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.279690 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.282807 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de309abf-bc67-416d-b0b0-1da8f0e952a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.299142 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.300367 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de309abf-bc67-416d-b0b0-1da8f0e952a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.320634 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.339380 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.342066 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a25551-3a54-4eb2-8e5a-e13134f2ea95-serving-cert\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.360100 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.379987 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.392705 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ce76625d-d545-4664-b365-97cd648cb7ab-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.400060 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.428911 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.431308 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b689e3e-c89a-42bc-b6df-180d587cc197-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.439770 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.460715 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.480558 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.494633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b689e3e-c89a-42bc-b6df-180d587cc197-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.499749 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.520495 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.525984 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9aa1836d-6217-4c0f-bb26-dfc40643305d-metrics-tls\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.540288 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.561445 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.580959 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.594234 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5b600621-f833-48ef-9d9d-04f6543d81ae-proxy-tls\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.599679 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.625977 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.630883 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/059e5978-1a83-45f0-876a-c497fa2365b6-trusted-ca\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.641277 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.653381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/059e5978-1a83-45f0-876a-c497fa2365b6-metrics-tls\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.659729 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.679896 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.721049 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.741537 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.760055 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.780365 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.800661 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.815010 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/99337d25-bf3b-4740-9e30-8ef50cecd6cd-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.820666 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.841126 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.855330 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-serving-cert\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.859932 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.871264 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-config\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.881078 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.900151 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.920436 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.941580 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.960245 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 14 07:53:27 crc kubenswrapper[5002]: I1014 07:53:27.980593 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.001175 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.015490 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-webhook-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.015699 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/007477a4-2f7b-46a0-9254-fe1da6f900f4-apiservice-cert\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.019820 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.041233 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.061394 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.080436 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.094607 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-srv-cert\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.100223 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.118163 5002 request.go:700] Waited for 1.002906617s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dsigning-cabundle&limit=500&resourceVersion=0 Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.120286 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.131872 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-cabundle\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.140447 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.160145 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.170758 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.181575 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.191673 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-config\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.200580 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.214756 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/77e57d6f-a931-4fee-a933-2abb2cbe22d3-signing-key\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.220434 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.240273 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249868 5002 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249891 5002 secret.go:188] Couldn't get secret openshift-ingress/router-metrics-certs-default: failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249934 5002 secret.go:188] Couldn't get secret openshift-ingress/router-certs-default: failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249961 5002 secret.go:188] Couldn't get secret openshift-ingress/router-stats-default: failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249991 5002 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.249972 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images podName:1a0b16d3-a2ae-4c97-a0cf-e3ba76822395 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.749942438 +0000 UTC m=+141.731181940 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images") pod "machine-config-operator-74547568cd-5s65x" (UID: "1a0b16d3-a2ae-4c97-a0cf-e3ba76822395") : failed to sync configmap cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250073 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs podName:9f6b37d8-9ab1-41e6-a511-3116e9441ab2 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.750045821 +0000 UTC m=+141.731285303 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs") pod "router-default-5444994796-dndgk" (UID: "9f6b37d8-9ab1-41e6-a511-3116e9441ab2") : failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250097 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls podName:1a0b16d3-a2ae-4c97-a0cf-e3ba76822395 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.750085172 +0000 UTC m=+141.731324654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls") pod "machine-config-operator-74547568cd-5s65x" (UID: "1a0b16d3-a2ae-4c97-a0cf-e3ba76822395") : failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250098 5002 configmap.go:193] Couldn't get configMap openshift-ingress/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250123 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth podName:9f6b37d8-9ab1-41e6-a511-3116e9441ab2 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.750109074 +0000 UTC m=+141.731348566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "stats-auth" (UniqueName: "kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth") pod "router-default-5444994796-dndgk" (UID: "9f6b37d8-9ab1-41e6-a511-3116e9441ab2") : failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250201 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle podName:9f6b37d8-9ab1-41e6-a511-3116e9441ab2 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.750171786 +0000 UTC m=+141.731411268 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle") pod "router-default-5444994796-dndgk" (UID: "9f6b37d8-9ab1-41e6-a511-3116e9441ab2") : failed to sync configmap cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.250278 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate podName:9f6b37d8-9ab1-41e6-a511-3116e9441ab2 nodeName:}" failed. No retries permitted until 2025-10-14 07:53:28.750258738 +0000 UTC m=+141.731498310 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-certificate" (UniqueName: "kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate") pod "router-default-5444994796-dndgk" (UID: "9f6b37d8-9ab1-41e6-a511-3116e9441ab2") : failed to sync secret cache: timed out waiting for the condition Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.259649 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.280236 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.299982 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.319996 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.341107 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.360476 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.380763 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.400547 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.440761 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.460115 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.480310 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.520896 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.528457 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jsrt\" (UniqueName: \"kubernetes.io/projected/841fda0e-a272-490f-81ab-dc83e48b2201-kube-api-access-4jsrt\") pod \"authentication-operator-69f744f599-ml5x5\" (UID: \"841fda0e-a272-490f-81ab-dc83e48b2201\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.540442 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.560881 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.580105 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 14 07:53:28 crc kubenswrapper[5002]: E1014 07:53:28.619288 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.620105 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.628147 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qc94\" (UniqueName: \"kubernetes.io/projected/f9bcc50b-a9bb-4ccf-bb92-42718291689d-kube-api-access-9qc94\") pod \"openshift-apiserver-operator-796bbdcf4f-xn8pn\" (UID: \"f9bcc50b-a9bb-4ccf-bb92-42718291689d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.640879 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.689051 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8cbb\" (UniqueName: \"kubernetes.io/projected/6b840956-6900-4c7c-89ce-622ddf455cf6-kube-api-access-g8cbb\") pod \"control-plane-machine-set-operator-78cbb6b69f-9gbqd\" (UID: \"6b840956-6900-4c7c-89ce-622ddf455cf6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.707712 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thn47\" (UniqueName: \"kubernetes.io/projected/3fdd526f-086e-4fb1-8a38-d297ae1843e6-kube-api-access-thn47\") pod \"apiserver-7bbb656c7d-h4xjg\" (UID: \"3fdd526f-086e-4fb1-8a38-d297ae1843e6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.716361 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.734164 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmgs\" (UniqueName: \"kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs\") pod \"controller-manager-879f6c89f-hbt2w\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.752164 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkbmz\" (UniqueName: \"kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz\") pod \"route-controller-manager-6576b87f9c-mh8hc\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.761047 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.774715 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ndd5\" (UniqueName: \"kubernetes.io/projected/bc3a04d3-b88a-4a0a-84b8-618cdd10d617-kube-api-access-9ndd5\") pod \"machine-approver-56656f9798-2njf4\" (UID: \"bc3a04d3-b88a-4a0a-84b8-618cdd10d617\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.778153 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.778343 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.778417 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.778630 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.779283 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.779516 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.779918 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-images\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.781313 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-service-ca-bundle\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.784520 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-metrics-certs\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.784826 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-stats-auth\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.786648 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-proxy-tls\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.786943 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-default-certificate\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.801063 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.802667 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.815362 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.815968 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx4wp\" (UniqueName: \"kubernetes.io/projected/0b5fc9a0-1cd3-479a-b0b7-6a9e907128be-kube-api-access-wx4wp\") pod \"cluster-samples-operator-665b6dd947-b7sbl\" (UID: \"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.820591 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.853352 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.860566 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.901543 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msvb4\" (UniqueName: \"kubernetes.io/projected/71022b1b-51c8-4c8c-8d57-dc220ebf1b05-kube-api-access-msvb4\") pod \"machine-api-operator-5694c8668f-hjt8h\" (UID: \"71022b1b-51c8-4c8c-8d57-dc220ebf1b05\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.904376 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.917153 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt7rl\" (UniqueName: \"kubernetes.io/projected/01c76e0a-94db-4229-8201-dfb777a5ccd0-kube-api-access-qt7rl\") pod \"etcd-operator-b45778765-q9jk6\" (UID: \"01c76e0a-94db-4229-8201-dfb777a5ccd0\") " pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.924890 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.945607 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npp2w\" (UniqueName: \"kubernetes.io/projected/433fc6e4-cadb-45b7-964b-cd0a82c09897-kube-api-access-npp2w\") pod \"console-operator-58897d9998-dg9pb\" (UID: \"433fc6e4-cadb-45b7-964b-cd0a82c09897\") " pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.960446 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qw86\" (UniqueName: \"kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86\") pod \"oauth-openshift-558db77b4-grb97\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.978477 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4m5d\" (UniqueName: \"kubernetes.io/projected/7ea0cd04-9117-4160-bb93-e211cf1cba70-kube-api-access-b4m5d\") pod \"olm-operator-6b444d44fb-p9trk\" (UID: \"7ea0cd04-9117-4160-bb93-e211cf1cba70\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.982553 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.989606 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" Oct 14 07:53:28 crc kubenswrapper[5002]: I1014 07:53:28.995572 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.000447 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.003082 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.004269 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gnm\" (UniqueName: \"kubernetes.io/projected/42350c04-b619-45e9-8567-8f71d4c95e46-kube-api-access-n9gnm\") pod \"apiserver-76f77b778f-pd9r6\" (UID: \"42350c04-b619-45e9-8567-8f71d4c95e46\") " pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.021860 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.041141 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.042475 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.043328 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:29 crc kubenswrapper[5002]: W1014 07:53:29.052749 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc3a04d3_b88a_4a0a_84b8_618cdd10d617.slice/crio-9807279fa697cc08584e232490e59b6cc818bf5fc853335cfeadd230480b2b6a WatchSource:0}: Error finding container 9807279fa697cc08584e232490e59b6cc818bf5fc853335cfeadd230480b2b6a: Status 404 returned error can't find the container with id 9807279fa697cc08584e232490e59b6cc818bf5fc853335cfeadd230480b2b6a Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.062274 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ml5x5"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.077854 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lqn9\" (UniqueName: \"kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9\") pod \"console-f9d7485db-sn4pd\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.081876 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.083588 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.103091 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.119725 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.134100 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.139297 5002 request.go:700] Waited for 1.955936414s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.140647 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.143544 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.154060 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.160572 5002 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.168759 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.174569 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:29 crc kubenswrapper[5002]: W1014 07:53:29.176953 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fdd526f_086e_4fb1_8a38_d297ae1843e6.slice/crio-9bd4bf6897412b7f9c3f9886ef0ffd8f6aa861f66d5aa17d3af649095024a325 WatchSource:0}: Error finding container 9bd4bf6897412b7f9c3f9886ef0ffd8f6aa861f66d5aa17d3af649095024a325: Status 404 returned error can't find the container with id 9bd4bf6897412b7f9c3f9886ef0ffd8f6aa861f66d5aa17d3af649095024a325 Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.182168 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.192815 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.203090 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.203967 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.223886 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.240427 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.263137 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.307230 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.329508 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4726\" (UniqueName: \"kubernetes.io/projected/ce76625d-d545-4664-b365-97cd648cb7ab-kube-api-access-l4726\") pod \"multus-admission-controller-857f4d67dd-fv4sf\" (UID: \"ce76625d-d545-4664-b365-97cd648cb7ab\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.344604 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qgb2\" (UniqueName: \"kubernetes.io/projected/9aa1836d-6217-4c0f-bb26-dfc40643305d-kube-api-access-2qgb2\") pod \"dns-operator-744455d44c-zjgv5\" (UID: \"9aa1836d-6217-4c0f-bb26-dfc40643305d\") " pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.349968 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.352986 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4smz\" (UniqueName: \"kubernetes.io/projected/77e57d6f-a931-4fee-a933-2abb2cbe22d3-kube-api-access-s4smz\") pod \"service-ca-9c57cc56f-wmg78\" (UID: \"77e57d6f-a931-4fee-a933-2abb2cbe22d3\") " pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.373969 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfzxt\" (UniqueName: \"kubernetes.io/projected/20a25551-3a54-4eb2-8e5a-e13134f2ea95-kube-api-access-qfzxt\") pod \"openshift-config-operator-7777fb866f-s656d\" (UID: \"20a25551-3a54-4eb2-8e5a-e13134f2ea95\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.385245 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.398460 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slhr5\" (UniqueName: \"kubernetes.io/projected/99337d25-bf3b-4740-9e30-8ef50cecd6cd-kube-api-access-slhr5\") pod \"package-server-manager-789f6589d5-5nc9n\" (UID: \"99337d25-bf3b-4740-9e30-8ef50cecd6cd\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.418984 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppcn8\" (UniqueName: \"kubernetes.io/projected/007477a4-2f7b-46a0-9254-fe1da6f900f4-kube-api-access-ppcn8\") pod \"packageserver-d55dfcdfc-jpw66\" (UID: \"007477a4-2f7b-46a0-9254-fe1da6f900f4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.423204 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.435115 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4hj5\" (UniqueName: \"kubernetes.io/projected/9f6b37d8-9ab1-41e6-a511-3116e9441ab2-kube-api-access-q4hj5\") pod \"router-default-5444994796-dndgk\" (UID: \"9f6b37d8-9ab1-41e6-a511-3116e9441ab2\") " pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.435432 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.436553 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.443906 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.454431 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a35c3bf5-d486-4502-9a31-0dc7848e4bd8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-tk88f\" (UID: \"a35c3bf5-d486-4502-9a31-0dc7848e4bd8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.455290 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.463144 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.468455 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:29 crc kubenswrapper[5002]: W1014 07:53:29.470121 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17bc792f_98b0_46b1_b95c_00f3ee5e40a5.slice/crio-1167936feb200e580244d138c63988345a1cb35d7555a19c1576867cc7a629af WatchSource:0}: Error finding container 1167936feb200e580244d138c63988345a1cb35d7555a19c1576867cc7a629af: Status 404 returned error can't find the container with id 1167936feb200e580244d138c63988345a1cb35d7555a19c1576867cc7a629af Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.475122 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgt7z\" (UniqueName: \"kubernetes.io/projected/059e5978-1a83-45f0-876a-c497fa2365b6-kube-api-access-dgt7z\") pod \"ingress-operator-5b745b69d9-dvnpb\" (UID: \"059e5978-1a83-45f0-876a-c497fa2365b6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.509456 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4nn\" (UniqueName: \"kubernetes.io/projected/a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84-kube-api-access-jl4nn\") pod \"service-ca-operator-777779d784-l9m77\" (UID: \"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.526473 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzss2\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-kube-api-access-nzss2\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.543771 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzjvs\" (UniqueName: \"kubernetes.io/projected/f6a3e3bd-467a-4523-89d3-0fb063ba2a94-kube-api-access-fzjvs\") pod \"catalog-operator-68c6474976-5g28n\" (UID: \"f6a3e3bd-467a-4523-89d3-0fb063ba2a94\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.553552 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" event={"ID":"f9bcc50b-a9bb-4ccf-bb92-42718291689d","Type":"ContainerStarted","Data":"c66cbca6e06a02c0ccf714ab1f1e66aa98825424229fd3a73abca3ecf6727f5f"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.554006 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pd9r6"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.554022 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" event={"ID":"f9bcc50b-a9bb-4ccf-bb92-42718291689d","Type":"ContainerStarted","Data":"dc6955efdffba7c756661eab7a6823e75f31efead073235ca197ca1d721b503d"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.555257 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.558962 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-q9jk6"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.565371 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v79pj\" (UniqueName: \"kubernetes.io/projected/1a0b16d3-a2ae-4c97-a0cf-e3ba76822395-kube-api-access-v79pj\") pod \"machine-config-operator-74547568cd-5s65x\" (UID: \"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.569144 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" event={"ID":"bc3a04d3-b88a-4a0a-84b8-618cdd10d617","Type":"ContainerStarted","Data":"65645a4d97edc1f3bbc61915aba953f159cd45a1f1fdde7fe6d9c4681774af3d"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.569182 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" event={"ID":"bc3a04d3-b88a-4a0a-84b8-618cdd10d617","Type":"ContainerStarted","Data":"9807279fa697cc08584e232490e59b6cc818bf5fc853335cfeadd230480b2b6a"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.570426 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" event={"ID":"17bc792f-98b0-46b1-b95c-00f3ee5e40a5","Type":"ContainerStarted","Data":"1167936feb200e580244d138c63988345a1cb35d7555a19c1576867cc7a629af"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.572859 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" event={"ID":"841fda0e-a272-490f-81ab-dc83e48b2201","Type":"ContainerStarted","Data":"1d132b2340bc306a8904c59e05367e6bf59ee89b7c52ce98eaf58c4e77d064eb"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.576267 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26k5p\" (UniqueName: \"kubernetes.io/projected/de309abf-bc67-416d-b0b0-1da8f0e952a2-kube-api-access-26k5p\") pod \"openshift-controller-manager-operator-756b6f6bc6-6pz4b\" (UID: \"de309abf-bc67-416d-b0b0-1da8f0e952a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.579098 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" event={"ID":"47b9ba60-7de6-4bd5-bb0f-edcf124193ad","Type":"ContainerStarted","Data":"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.579138 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" event={"ID":"47b9ba60-7de6-4bd5-bb0f-edcf124193ad","Type":"ContainerStarted","Data":"ab58f7a574edc35ca3602108e35435246e577800971724bcbf0c6baedf9db504"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.579402 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.581989 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" event={"ID":"6b840956-6900-4c7c-89ce-622ddf455cf6","Type":"ContainerStarted","Data":"7aa6e09cd92886ab8e6645624feeef7b2e7d86991fb6a981179a4293eed5f23d"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.587532 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" event={"ID":"3fdd526f-086e-4fb1-8a38-d297ae1843e6","Type":"ContainerStarted","Data":"9bd4bf6897412b7f9c3f9886ef0ffd8f6aa861f66d5aa17d3af649095024a325"} Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.596322 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b689e3e-c89a-42bc-b6df-180d587cc197-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-tf7xq\" (UID: \"5b689e3e-c89a-42bc-b6df-180d587cc197\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.598492 5002 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hbt2w container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.598533 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 14 07:53:29 crc kubenswrapper[5002]: W1014 07:53:29.600897 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f6b37d8_9ab1_41e6_a511_3116e9441ab2.slice/crio-c771903a8e84537b002b95a2ed6e20fd9fcd478ed60d5dc0afb8e632e71a79eb WatchSource:0}: Error finding container c771903a8e84537b002b95a2ed6e20fd9fcd478ed60d5dc0afb8e632e71a79eb: Status 404 returned error can't find the container with id c771903a8e84537b002b95a2ed6e20fd9fcd478ed60d5dc0afb8e632e71a79eb Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.619549 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hjt8h"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.626637 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wbq2\" (UniqueName: \"kubernetes.io/projected/89e82ba1-52a9-4c22-bc43-7d71bf835d2f-kube-api-access-8wbq2\") pod \"migrator-59844c95c7-4r89t\" (UID: \"89e82ba1-52a9-4c22-bc43-7d71bf835d2f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.635098 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.641252 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dg9pb"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.642725 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.647598 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v9dw\" (UniqueName: \"kubernetes.io/projected/5b600621-f833-48ef-9d9d-04f6543d81ae-kube-api-access-4v9dw\") pod \"machine-config-controller-84d6567774-sdc6q\" (UID: \"5b600621-f833-48ef-9d9d-04f6543d81ae\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.662062 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.669950 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2rjk\" (UniqueName: \"kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk\") pod \"marketplace-operator-79b997595-ndpjk\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.671356 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fv4sf"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.677182 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.690911 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.697819 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12829eb0-c7da-49ac-8350-ce7329fded6b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.697938 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.697994 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12829eb0-c7da-49ac-8350-ce7329fded6b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698019 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnvzx\" (UniqueName: \"kubernetes.io/projected/12829eb0-c7da-49ac-8350-ce7329fded6b-kube-api-access-tnvzx\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698052 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698076 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698113 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698144 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8fqj\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698174 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698214 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698241 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.698368 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" Oct 14 07:53:29 crc kubenswrapper[5002]: E1014 07:53:29.699515 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.199463457 +0000 UTC m=+143.180702909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.701816 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zjgv5"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.717007 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.742418 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.749710 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.777799 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.778385 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800054 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:29 crc kubenswrapper[5002]: E1014 07:53:29.800205 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.300182844 +0000 UTC m=+143.281422296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800348 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c8zk\" (UniqueName: \"kubernetes.io/projected/74ba69e9-c389-416e-aef3-9645524e6944-kube-api-access-7c8zk\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800444 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-config-volume\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800501 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800665 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800740 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9ccz\" (UniqueName: \"kubernetes.io/projected/3ccad502-97aa-44b3-bb4f-d981ec035d20-kube-api-access-f9ccz\") pod \"downloads-7954f5f757-xdlrk\" (UID: \"3ccad502-97aa-44b3-bb4f-d981ec035d20\") " pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.800762 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.801115 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f691612-f354-4962-875d-8ff20478a22d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: E1014 07:53:29.801215 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.301203562 +0000 UTC m=+143.282443014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.801658 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12829eb0-c7da-49ac-8350-ce7329fded6b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.801990 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802091 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802216 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-node-bootstrap-token\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802455 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-certs\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802577 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802699 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4zzv\" (UniqueName: \"kubernetes.io/projected/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-kube-api-access-p4zzv\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802861 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-registration-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802922 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12829eb0-c7da-49ac-8350-ce7329fded6b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.802940 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnvzx\" (UniqueName: \"kubernetes.io/projected/12829eb0-c7da-49ac-8350-ce7329fded6b-kube-api-access-tnvzx\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803033 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f691612-f354-4962-875d-8ff20478a22d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803100 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803121 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f691612-f354-4962-875d-8ff20478a22d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803160 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8570c9-12e7-488e-aeac-c6c566122868-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803237 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qmp2\" (UniqueName: \"kubernetes.io/projected/3566d84f-2e3b-4f11-ad84-6f33541eced7-kube-api-access-7qmp2\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803681 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-socket-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803719 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb4gq\" (UniqueName: \"kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.803772 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/74ba69e9-c389-416e-aef3-9645524e6944-cert\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.804489 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.804512 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12829eb0-c7da-49ac-8350-ce7329fded6b-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.804599 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njgtj\" (UniqueName: \"kubernetes.io/projected/b6350740-ff4e-484d-b816-848f96e98525-kube-api-access-njgtj\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.804818 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-csi-data-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.804863 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8570c9-12e7-488e-aeac-c6c566122868-config\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.805137 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.805571 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8fqj\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.805788 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-mountpoint-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.806171 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-plugins-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.806198 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-metrics-tls\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.806215 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a8570c9-12e7-488e-aeac-c6c566122868-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.807969 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.809287 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12829eb0-c7da-49ac-8350-ce7329fded6b-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.811015 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.826733 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.842407 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: W1014 07:53:29.853901 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6389e00b_ba44_426e_b07a_97e18189d9ef.slice/crio-bf1b657658355473476a75e25fb6ce6bd2541d5079dee274d837a81120005bd2 WatchSource:0}: Error finding container bf1b657658355473476a75e25fb6ce6bd2541d5079dee274d837a81120005bd2: Status 404 returned error can't find the container with id bf1b657658355473476a75e25fb6ce6bd2541d5079dee274d837a81120005bd2 Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.862022 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnvzx\" (UniqueName: \"kubernetes.io/projected/12829eb0-c7da-49ac-8350-ce7329fded6b-kube-api-access-tnvzx\") pod \"kube-storage-version-migrator-operator-b67b599dd-9httn\" (UID: \"12829eb0-c7da-49ac-8350-ce7329fded6b\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.887031 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8fqj\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.907615 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.907927 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.907963 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-node-bootstrap-token\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.907983 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-certs\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908008 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4zzv\" (UniqueName: \"kubernetes.io/projected/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-kube-api-access-p4zzv\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908034 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-registration-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908068 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f691612-f354-4962-875d-8ff20478a22d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908085 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f691612-f354-4962-875d-8ff20478a22d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908100 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8570c9-12e7-488e-aeac-c6c566122868-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908118 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qmp2\" (UniqueName: \"kubernetes.io/projected/3566d84f-2e3b-4f11-ad84-6f33541eced7-kube-api-access-7qmp2\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908135 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-socket-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908150 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb4gq\" (UniqueName: \"kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908164 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/74ba69e9-c389-416e-aef3-9645524e6944-cert\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.908182 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njgtj\" (UniqueName: \"kubernetes.io/projected/b6350740-ff4e-484d-b816-848f96e98525-kube-api-access-njgtj\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910061 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-csi-data-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910080 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8570c9-12e7-488e-aeac-c6c566122868-config\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910152 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-mountpoint-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910176 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-plugins-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910195 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-metrics-tls\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910210 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a8570c9-12e7-488e-aeac-c6c566122868-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910232 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c8zk\" (UniqueName: \"kubernetes.io/projected/74ba69e9-c389-416e-aef3-9645524e6944-kube-api-access-7c8zk\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910251 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-config-volume\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910279 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9ccz\" (UniqueName: \"kubernetes.io/projected/3ccad502-97aa-44b3-bb4f-d981ec035d20-kube-api-access-f9ccz\") pod \"downloads-7954f5f757-xdlrk\" (UID: \"3ccad502-97aa-44b3-bb4f-d981ec035d20\") " pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910293 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910354 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f691612-f354-4962-875d-8ff20478a22d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: E1014 07:53:29.910613 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.410593561 +0000 UTC m=+143.391833013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910785 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-registration-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.910965 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-socket-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.914979 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-csi-data-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.915416 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-mountpoint-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.915480 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3566d84f-2e3b-4f11-ad84-6f33541eced7-plugins-dir\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.916152 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a8570c9-12e7-488e-aeac-c6c566122868-config\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.916596 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-config-volume\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.917231 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f691612-f354-4962-875d-8ff20478a22d-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.918204 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.921530 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f"] Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.925396 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.934144 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/74ba69e9-c389-416e-aef3-9645524e6944-cert\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.937713 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-metrics-tls\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.946228 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.948682 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-certs\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.949285 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/b6350740-ff4e-484d-b816-848f96e98525-node-bootstrap-token\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.958965 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a8570c9-12e7-488e-aeac-c6c566122868-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.959152 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f691612-f354-4962-875d-8ff20478a22d-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.973607 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb4gq\" (UniqueName: \"kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq\") pod \"collect-profiles-29340465-h7nfm\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:29 crc kubenswrapper[5002]: I1014 07:53:29.975892 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f691612-f354-4962-875d-8ff20478a22d-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6wv7p\" (UID: \"4f691612-f354-4962-875d-8ff20478a22d\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.000305 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wmg78"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.004886 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.005774 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qmp2\" (UniqueName: \"kubernetes.io/projected/3566d84f-2e3b-4f11-ad84-6f33541eced7-kube-api-access-7qmp2\") pod \"csi-hostpathplugin-cnwsm\" (UID: \"3566d84f-2e3b-4f11-ad84-6f33541eced7\") " pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.013020 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.013438 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.513423589 +0000 UTC m=+143.494663041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.029669 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.032860 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.039036 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2a8570c9-12e7-488e-aeac-c6c566122868-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gwzw5\" (UID: \"2a8570c9-12e7-488e-aeac-c6c566122868\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.044804 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c8zk\" (UniqueName: \"kubernetes.io/projected/74ba69e9-c389-416e-aef3-9645524e6944-kube-api-access-7c8zk\") pod \"ingress-canary-s6s5j\" (UID: \"74ba69e9-c389-416e-aef3-9645524e6944\") " pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:30 crc kubenswrapper[5002]: W1014 07:53:30.061491 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99337d25_bf3b_4740_9e30_8ef50cecd6cd.slice/crio-d2b1abae7bf03af553ce9e491e706bf6f1e65c448eafb2aedbc0b0fbfa1ba1d1 WatchSource:0}: Error finding container d2b1abae7bf03af553ce9e491e706bf6f1e65c448eafb2aedbc0b0fbfa1ba1d1: Status 404 returned error can't find the container with id d2b1abae7bf03af553ce9e491e706bf6f1e65c448eafb2aedbc0b0fbfa1ba1d1 Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.062204 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9ccz\" (UniqueName: \"kubernetes.io/projected/3ccad502-97aa-44b3-bb4f-d981ec035d20-kube-api-access-f9ccz\") pod \"downloads-7954f5f757-xdlrk\" (UID: \"3ccad502-97aa-44b3-bb4f-d981ec035d20\") " pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.076143 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njgtj\" (UniqueName: \"kubernetes.io/projected/b6350740-ff4e-484d-b816-848f96e98525-kube-api-access-njgtj\") pod \"machine-config-server-l8k2s\" (UID: \"b6350740-ff4e-484d-b816-848f96e98525\") " pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.087240 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.093462 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.098641 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.100515 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4zzv\" (UniqueName: \"kubernetes.io/projected/2d0b4cdd-b5c1-46ad-a1db-1780877eb35a-kube-api-access-p4zzv\") pod \"dns-default-rjkc7\" (UID: \"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a\") " pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.106035 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.115016 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-l8k2s" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.116865 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.116914 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.616897563 +0000 UTC m=+143.598137015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.117215 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.117564 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.617555701 +0000 UTC m=+143.598795153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.119310 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-s6s5j" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.152568 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.152589 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.176261 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-l9m77"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.196788 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.211689 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-s656d"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.218820 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.219239 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.719223976 +0000 UTC m=+143.700463428 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.321132 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.321430 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.821417994 +0000 UTC m=+143.802657446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.422144 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.422308 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.922283516 +0000 UTC m=+143.903522968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.422761 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.423118 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:30.923096578 +0000 UTC m=+143.904336110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.449418 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.487471 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.507173 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.523430 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.523520 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.023499257 +0000 UTC m=+144.004738709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.523958 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.524342 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.024328351 +0000 UTC m=+144.005567803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.588434 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.593115 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.617900 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" event={"ID":"5b689e3e-c89a-42bc-b6df-180d587cc197","Type":"ContainerStarted","Data":"d0338f7ac6a0dce50672b60838fd3051b060bc869027b16fabedd8b6fd510d6f"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.627568 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.628008 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.127993241 +0000 UTC m=+144.109232693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.632643 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.672129 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sn4pd" event={"ID":"6389e00b-ba44-426e-b07a-97e18189d9ef","Type":"ContainerStarted","Data":"bf1b657658355473476a75e25fb6ce6bd2541d5079dee274d837a81120005bd2"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.674757 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.675926 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.707335 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" event={"ID":"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc","Type":"ContainerStarted","Data":"a322e84f397cf4ce0576275f3e644ea0b9d273c8c1bbbd2b583a270e135ab3a1"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.733383 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.733675 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.233663666 +0000 UTC m=+144.214903118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: W1014 07:53:30.765179 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd63d35e_85fa_44ae_9389_28b816cffae3.slice/crio-bb693d7c92d1c067952113bb2be8b9fb76ca5b195ecda44a58877d0a9d7e6bf6 WatchSource:0}: Error finding container bb693d7c92d1c067952113bb2be8b9fb76ca5b195ecda44a58877d0a9d7e6bf6: Status 404 returned error can't find the container with id bb693d7c92d1c067952113bb2be8b9fb76ca5b195ecda44a58877d0a9d7e6bf6 Oct 14 07:53:30 crc kubenswrapper[5002]: W1014 07:53:30.784374 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6a3e3bd_467a_4523_89d3_0fb063ba2a94.slice/crio-3b57cb0aeb0e065e22dc83d9b22dbb9fe5c17f8c741d67c5aa4b3f04a0ade2ae WatchSource:0}: Error finding container 3b57cb0aeb0e065e22dc83d9b22dbb9fe5c17f8c741d67c5aa4b3f04a0ade2ae: Status 404 returned error can't find the container with id 3b57cb0aeb0e065e22dc83d9b22dbb9fe5c17f8c741d67c5aa4b3f04a0ade2ae Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.786159 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" event={"ID":"6b840956-6900-4c7c-89ce-622ddf455cf6","Type":"ContainerStarted","Data":"43d342c76eafb25e49f64c376665d2a5f020cd143083bcacec06ccd4fc663ff8"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.797351 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" event={"ID":"7ea0cd04-9117-4160-bb93-e211cf1cba70","Type":"ContainerStarted","Data":"8580d7146adde333a19b7d22240ba979fa8f9e2cb49cbdb2430434eef4ec8805"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.806128 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.806178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" event={"ID":"7ea0cd04-9117-4160-bb93-e211cf1cba70","Type":"ContainerStarted","Data":"36f36d8801368ffe50c34bb2fcc2299e9d65f91e2553451f1e49ad308532df93"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.814463 5002 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p9trk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.814509 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" podUID="7ea0cd04-9117-4160-bb93-e211cf1cba70" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.816592 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" event={"ID":"433fc6e4-cadb-45b7-964b-cd0a82c09897","Type":"ContainerStarted","Data":"40e006dab17c4aa398069227ceb0e81c32fc05f32cae93c7e795abc000755e08"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.816635 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" event={"ID":"433fc6e4-cadb-45b7-964b-cd0a82c09897","Type":"ContainerStarted","Data":"f858c3e260cfc0d0fd6c3ff4e73588d277ec652bb147de12a480253fefb0d305"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.817378 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.831555 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" event={"ID":"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be","Type":"ContainerStarted","Data":"6ea6ab50857558e881c4599871deb7779e3b43075a035e8cf20ae202f7d6f20d"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.831801 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" event={"ID":"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be","Type":"ContainerStarted","Data":"5a7fb64fb4761981dcf781c3004a04772a88fda9823da01bc46ea999f881297a"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.836228 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.837794 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.337700717 +0000 UTC m=+144.318940169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.847368 5002 patch_prober.go:28] interesting pod/console-operator-58897d9998-dg9pb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.847421 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" podUID="433fc6e4-cadb-45b7-964b-cd0a82c09897" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.861276 5002 generic.go:334] "Generic (PLEG): container finished" podID="3fdd526f-086e-4fb1-8a38-d297ae1843e6" containerID="5c3d8be1375799b9a086822b69ec7fb9f698e7c642298afec028cbf194229e93" exitCode=0 Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.861341 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" event={"ID":"3fdd526f-086e-4fb1-8a38-d297ae1843e6","Type":"ContainerDied","Data":"5c3d8be1375799b9a086822b69ec7fb9f698e7c642298afec028cbf194229e93"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.873718 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" podStartSLOduration=116.873582256 podStartE2EDuration="1m56.873582256s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:30.873583286 +0000 UTC m=+143.854822748" watchObservedRunningTime="2025-10-14 07:53:30.873582256 +0000 UTC m=+143.854821708" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.875582 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" event={"ID":"99337d25-bf3b-4740-9e30-8ef50cecd6cd","Type":"ContainerStarted","Data":"d2b1abae7bf03af553ce9e491e706bf6f1e65c448eafb2aedbc0b0fbfa1ba1d1"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.878438 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" event={"ID":"01c76e0a-94db-4229-8201-dfb777a5ccd0","Type":"ContainerStarted","Data":"f1a4e0398755674eee3ee11ce438d92077448ce3a2015ddeffc1c5c4dd690a0b"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.878464 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" event={"ID":"01c76e0a-94db-4229-8201-dfb777a5ccd0","Type":"ContainerStarted","Data":"39ff2b5960c4300049a1092a98069ccf00a9245f68e8eab47b29a6e7b145cf2e"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.905442 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" event={"ID":"ce76625d-d545-4664-b365-97cd648cb7ab","Type":"ContainerStarted","Data":"5f5524e2b0458d0384b204123aa053dc428f999286e525d9acda69179824aa50"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.905484 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" event={"ID":"ce76625d-d545-4664-b365-97cd648cb7ab","Type":"ContainerStarted","Data":"754cc58f5c9ed7c0abe59f168557f4939b9f36cb64226172f158b6b9ef1cbd39"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.910618 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" event={"ID":"841fda0e-a272-490f-81ab-dc83e48b2201","Type":"ContainerStarted","Data":"48420a3d224c674324952ef6cd0f32097f6b354ba8cf76fff909a439eda492d0"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.920324 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5"] Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.920428 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9gbqd" podStartSLOduration=116.920412002 podStartE2EDuration="1m56.920412002s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:30.910193967 +0000 UTC m=+143.891433419" watchObservedRunningTime="2025-10-14 07:53:30.920412002 +0000 UTC m=+143.901651454" Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.922804 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" event={"ID":"5b600621-f833-48ef-9d9d-04f6543d81ae","Type":"ContainerStarted","Data":"abc17020fd2bf942f9118f406c40294c37d635f810b11a5a0ca8bec8acd425af"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.932538 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" event={"ID":"9aa1836d-6217-4c0f-bb26-dfc40643305d","Type":"ContainerStarted","Data":"ffef6c6873c8cf33e8dcf8c2768d6958cd200ce185af159aa4de028da77a762d"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.935902 5002 generic.go:334] "Generic (PLEG): container finished" podID="42350c04-b619-45e9-8567-8f71d4c95e46" containerID="6221c10eecb741de000a81e9ee4ec8d63dcb6124d1757f3014c24ae077b14447" exitCode=0 Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.935965 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" event={"ID":"42350c04-b619-45e9-8567-8f71d4c95e46","Type":"ContainerDied","Data":"6221c10eecb741de000a81e9ee4ec8d63dcb6124d1757f3014c24ae077b14447"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.935992 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" event={"ID":"42350c04-b619-45e9-8567-8f71d4c95e46","Type":"ContainerStarted","Data":"ea023b29db6b10dc1a9f980fb9e5ceb0c06d87ed23e7abad9ab00a37e38291c6"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.939658 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:30 crc kubenswrapper[5002]: E1014 07:53:30.941013 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.440999316 +0000 UTC m=+144.422238768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.974030 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" event={"ID":"bc3a04d3-b88a-4a0a-84b8-618cdd10d617","Type":"ContainerStarted","Data":"e8707bbc162a30ef3bacb671c635c3661ff17ae1a8af0e00f371f2d425fda62d"} Oct 14 07:53:30 crc kubenswrapper[5002]: I1014 07:53:30.999280 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" event={"ID":"de309abf-bc67-416d-b0b0-1da8f0e952a2","Type":"ContainerStarted","Data":"077a2474406c71d1cb32c48a6f70ea40bc051bc7cbb732c0fd1ba0dba9f444a8"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.009746 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" event={"ID":"71022b1b-51c8-4c8c-8d57-dc220ebf1b05","Type":"ContainerStarted","Data":"3063b4474465a6418e42082582ea224a824b09c015218e5fb374f2cfb41cdb40"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.009785 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" event={"ID":"71022b1b-51c8-4c8c-8d57-dc220ebf1b05","Type":"ContainerStarted","Data":"d10d25660476cc1608a6ae964d3597d243068f0601821fd085472ddb2c26cdd9"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.010897 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" event={"ID":"20a25551-3a54-4eb2-8e5a-e13134f2ea95","Type":"ContainerStarted","Data":"17a08a892da91e42055a96b434a8b706423cf0fb7d171bdf76860efbf1343741"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.017525 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" event={"ID":"a35c3bf5-d486-4502-9a31-0dc7848e4bd8","Type":"ContainerStarted","Data":"5ed03007c88e4854e5147f7deb92b1e208196691ca9db4d434c5aa75518ba769"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.020861 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" event={"ID":"77e57d6f-a931-4fee-a933-2abb2cbe22d3","Type":"ContainerStarted","Data":"9c8e6107997bfc38980b5d81c2c8e98c9df19970cb17eb56bc6e50bb2ff6796e"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.024411 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" event={"ID":"007477a4-2f7b-46a0-9254-fe1da6f900f4","Type":"ContainerStarted","Data":"31b3dae13c18216eb8b4397309b709de31845b4f24043dd41838e7d639dd99aa"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.026274 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" event={"ID":"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84","Type":"ContainerStarted","Data":"767f56ae6c0e32a0eff9a503b71d77eb3e910a3ba63b515adc9f3d50e1c27351"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.041332 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.041718 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.541701133 +0000 UTC m=+144.522940585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.059449 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.117441 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xn8pn" podStartSLOduration=118.117427184 podStartE2EDuration="1m58.117427184s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:31.115209262 +0000 UTC m=+144.096448714" watchObservedRunningTime="2025-10-14 07:53:31.117427184 +0000 UTC m=+144.098666636" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.131469 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" event={"ID":"17bc792f-98b0-46b1-b95c-00f3ee5e40a5","Type":"ContainerStarted","Data":"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.131693 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.143012 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.143348 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.643337806 +0000 UTC m=+144.624577258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.149504 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dndgk" event={"ID":"9f6b37d8-9ab1-41e6-a511-3116e9441ab2","Type":"ContainerStarted","Data":"c56d192238b6dd4872ff701a6717821cbfe1c70771865013c5fb327b92e8ea03"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.149544 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dndgk" event={"ID":"9f6b37d8-9ab1-41e6-a511-3116e9441ab2","Type":"ContainerStarted","Data":"c771903a8e84537b002b95a2ed6e20fd9fcd478ed60d5dc0afb8e632e71a79eb"} Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.165011 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.243893 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.259948 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.759908566 +0000 UTC m=+144.741148018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.316404 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" podStartSLOduration=117.31638803 podStartE2EDuration="1m57.31638803s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:31.315218908 +0000 UTC m=+144.296458380" watchObservedRunningTime="2025-10-14 07:53:31.31638803 +0000 UTC m=+144.297627472" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.361294 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.365007 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.864990765 +0000 UTC m=+144.846230217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.462948 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.463427 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:31.963412378 +0000 UTC m=+144.944651830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.475235 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.486630 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:31 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:31 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:31 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.486677 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.496425 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.564773 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.565209 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.065197336 +0000 UTC m=+145.046436788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.618809 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.628112 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xdlrk"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.668271 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.668579 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.168564448 +0000 UTC m=+145.149803900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.753678 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rjkc7"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.754013 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cnwsm"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.770146 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.770687 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.270673204 +0000 UTC m=+145.251912656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.808821 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-dndgk" podStartSLOduration=117.808807767 podStartE2EDuration="1m57.808807767s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:31.756783617 +0000 UTC m=+144.738023079" watchObservedRunningTime="2025-10-14 07:53:31.808807767 +0000 UTC m=+144.790047219" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.831503 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-s6s5j"] Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.872333 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.872567 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.372553435 +0000 UTC m=+145.353792887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: W1014 07:53:31.933588 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3566d84f_2e3b_4f11_ad84_6f33541eced7.slice/crio-5f452232750cb659da6e718683741cfa11464192ebda94a7ac1825f839505eeb WatchSource:0}: Error finding container 5f452232750cb659da6e718683741cfa11464192ebda94a7ac1825f839505eeb: Status 404 returned error can't find the container with id 5f452232750cb659da6e718683741cfa11464192ebda94a7ac1825f839505eeb Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.973099 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:31 crc kubenswrapper[5002]: E1014 07:53:31.973358 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.473348284 +0000 UTC m=+145.454587736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.979016 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2njf4" podStartSLOduration=118.978999992 podStartE2EDuration="1m58.978999992s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:31.950124087 +0000 UTC m=+144.931363539" watchObservedRunningTime="2025-10-14 07:53:31.978999992 +0000 UTC m=+144.960239444" Oct 14 07:53:31 crc kubenswrapper[5002]: I1014 07:53:31.990038 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-q9jk6" podStartSLOduration=117.990020279 podStartE2EDuration="1m57.990020279s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:31.978626361 +0000 UTC m=+144.959865813" watchObservedRunningTime="2025-10-14 07:53:31.990020279 +0000 UTC m=+144.971259721" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.076524 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.077005 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.576987644 +0000 UTC m=+145.558227096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.099153 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ml5x5" podStartSLOduration=119.099137571 podStartE2EDuration="1m59.099137571s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.09732516 +0000 UTC m=+145.078564622" watchObservedRunningTime="2025-10-14 07:53:32.099137571 +0000 UTC m=+145.080377023" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.127957 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" podStartSLOduration=119.127941414 podStartE2EDuration="1m59.127941414s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.127472991 +0000 UTC m=+145.108712453" watchObservedRunningTime="2025-10-14 07:53:32.127941414 +0000 UTC m=+145.109180866" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.178780 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.179082 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.679071499 +0000 UTC m=+145.660310951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.183680 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" event={"ID":"dd63d35e-85fa-44ae-9389-28b816cffae3","Type":"ContainerStarted","Data":"bb693d7c92d1c067952113bb2be8b9fb76ca5b195ecda44a58877d0a9d7e6bf6"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.200597 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" event={"ID":"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be","Type":"ContainerStarted","Data":"2f40839cef765f27be536293eccb41c085941e8510cd912359ad01009d5ffc6c"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.206892 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" event={"ID":"007477a4-2f7b-46a0-9254-fe1da6f900f4","Type":"ContainerStarted","Data":"fb29ae2550a172e72a3301993dd20479a326f7613ac7f46d240513cfef650cf7"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.208590 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.208647 5002 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jpw66 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" start-of-body= Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.208672 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" podUID="007477a4-2f7b-46a0-9254-fe1da6f900f4" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.212427 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sn4pd" event={"ID":"6389e00b-ba44-426e-b07a-97e18189d9ef","Type":"ContainerStarted","Data":"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.233436 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" podStartSLOduration=118.233420674 podStartE2EDuration="1m58.233420674s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.231890711 +0000 UTC m=+145.213130193" watchObservedRunningTime="2025-10-14 07:53:32.233420674 +0000 UTC m=+145.214660126" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.238744 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" event={"ID":"89e82ba1-52a9-4c22-bc43-7d71bf835d2f","Type":"ContainerStarted","Data":"537342869dded4f3c31326f3d126b347d0fd2a38ccccf20c12ca931ed970cfa8"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.247927 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xdlrk" event={"ID":"3ccad502-97aa-44b3-bb4f-d981ec035d20","Type":"ContainerStarted","Data":"d48f7a5903f1a70a0331644f844676d3642911ce64b45951a3f4ebca29a96a93"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.251145 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l8k2s" event={"ID":"b6350740-ff4e-484d-b816-848f96e98525","Type":"ContainerStarted","Data":"0d35a169eb4d1af125d50354f38c7678ccdc7288d0cd8da4c313b9c3f289d068"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.253293 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" event={"ID":"f6a3e3bd-467a-4523-89d3-0fb063ba2a94","Type":"ContainerStarted","Data":"3b57cb0aeb0e065e22dc83d9b22dbb9fe5c17f8c741d67c5aa4b3f04a0ade2ae"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.255171 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" event={"ID":"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395","Type":"ContainerStarted","Data":"997e500985f6d305afc77e0309e2c652f9f0196ebdf8f0d4888982cdf03141d7"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.256206 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s6s5j" event={"ID":"74ba69e9-c389-416e-aef3-9645524e6944","Type":"ContainerStarted","Data":"d1090e262709f5023b52823eb80f83d5755ef90776421f7c6b688b7f27168f8f"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.258510 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" event={"ID":"9aa1836d-6217-4c0f-bb26-dfc40643305d","Type":"ContainerStarted","Data":"b65ce1e298fea40226001d87e9581c227e7043b639639578eace08d6eb6364a0"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.259529 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" event={"ID":"2a8570c9-12e7-488e-aeac-c6c566122868","Type":"ContainerStarted","Data":"afa48aaadfc8069ccbbcaff365e92ee9d9cc2dba0b5a984ef0830fae187cdaec"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.260531 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rjkc7" event={"ID":"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a","Type":"ContainerStarted","Data":"4b239333982627eefa66f6eabc714dc3de64b52cef5049da0c7658b71c354654"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.268733 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" event={"ID":"8f63d16a-af72-4d47-828d-8ac2532a6196","Type":"ContainerStarted","Data":"b9a0996b422906051ce44467ea563c302798b7cc5323f81db3bdefc5102bbe23"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.270110 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" event={"ID":"77e57d6f-a931-4fee-a933-2abb2cbe22d3","Type":"ContainerStarted","Data":"9c34bc2313eea21ba85c5da820c3b47d5da206e8f39938a39ff70c773c50ef47"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.282127 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.282412 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.782393779 +0000 UTC m=+145.763633231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.282916 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.283185 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.783174051 +0000 UTC m=+145.764413503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.284358 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" event={"ID":"71022b1b-51c8-4c8c-8d57-dc220ebf1b05","Type":"ContainerStarted","Data":"2883c2cef02d6007a33b3e68a74f3d56d29dff5b8f6761d61f84710006e6277b"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.292064 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" event={"ID":"3566d84f-2e3b-4f11-ad84-6f33541eced7","Type":"ContainerStarted","Data":"5f452232750cb659da6e718683741cfa11464192ebda94a7ac1825f839505eeb"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.314198 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" podStartSLOduration=118.314181765 podStartE2EDuration="1m58.314181765s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.281614318 +0000 UTC m=+145.262853780" watchObservedRunningTime="2025-10-14 07:53:32.314181765 +0000 UTC m=+145.295421217" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.318230 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" event={"ID":"99337d25-bf3b-4740-9e30-8ef50cecd6cd","Type":"ContainerStarted","Data":"472a1c9b78108aad3c3eb0f0984fc9535d17ae9bf2ea7f60f375344370952e69"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.333423 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" event={"ID":"20a25551-3a54-4eb2-8e5a-e13134f2ea95","Type":"ContainerStarted","Data":"9470f548a3aee49349a4b058bf82ee0849782e5a67300879407d8423a5765c43"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.337433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" event={"ID":"12829eb0-c7da-49ac-8350-ce7329fded6b","Type":"ContainerStarted","Data":"3f1b84f7be797c4d7ce8b62f7e80516e7716fd4f8ce583f6fa1e822b68bc5440"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.351870 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" event={"ID":"059e5978-1a83-45f0-876a-c497fa2365b6","Type":"ContainerStarted","Data":"8fed3add44b8760c865110cc9fa6e674b3f708273f9e8a7298d290a3675d876f"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.352182 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" podStartSLOduration=119.352172934 podStartE2EDuration="1m59.352172934s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.314794273 +0000 UTC m=+145.296033725" watchObservedRunningTime="2025-10-14 07:53:32.352172934 +0000 UTC m=+145.333412386" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.362148 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" event={"ID":"4f691612-f354-4962-875d-8ff20478a22d","Type":"ContainerStarted","Data":"0ab86c19f38f1f3781803541955897b180fddfda90035c0faa2806c1196fef86"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.379555 5002 patch_prober.go:28] interesting pod/console-operator-58897d9998-dg9pb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.379766 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" podUID="433fc6e4-cadb-45b7-964b-cd0a82c09897" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.8:8443/readyz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.379905 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" event={"ID":"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc","Type":"ContainerStarted","Data":"b7046fde711390b84f56d9e2f98b3e9f7c2abadb0c5c1ebea346d817d34431dd"} Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.384094 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.384506 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.884489195 +0000 UTC m=+145.865728647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.393512 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wmg78" podStartSLOduration=118.393496977 podStartE2EDuration="1m58.393496977s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.392688724 +0000 UTC m=+145.373928176" watchObservedRunningTime="2025-10-14 07:53:32.393496977 +0000 UTC m=+145.374736429" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.395098 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sn4pd" podStartSLOduration=119.395093001 podStartE2EDuration="1m59.395093001s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.355094946 +0000 UTC m=+145.336334408" watchObservedRunningTime="2025-10-14 07:53:32.395093001 +0000 UTC m=+145.376332453" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.406708 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9trk" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.435514 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hjt8h" podStartSLOduration=118.435500208 podStartE2EDuration="1m58.435500208s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.434037927 +0000 UTC m=+145.415277379" watchObservedRunningTime="2025-10-14 07:53:32.435500208 +0000 UTC m=+145.416739660" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.492126 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.492946 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:32 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:32 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:32 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.492993 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.497650 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:32.997635959 +0000 UTC m=+145.978875411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.597090 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.597184 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.097164584 +0000 UTC m=+146.078404036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.597317 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.597670 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.097662428 +0000 UTC m=+146.078901880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.698414 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.698873 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.198853159 +0000 UTC m=+146.180092611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.698919 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.699280 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.19926825 +0000 UTC m=+146.180507702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.810995 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.811588 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.31157099 +0000 UTC m=+146.292810442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:32 crc kubenswrapper[5002]: I1014 07:53:32.912584 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:32 crc kubenswrapper[5002]: E1014 07:53:32.912961 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.412950297 +0000 UTC m=+146.394189749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.014171 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.014479 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.514464896 +0000 UTC m=+146.495704348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.115679 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.116037 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.616025908 +0000 UTC m=+146.597265360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.216617 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.216750 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.716722025 +0000 UTC m=+146.697961477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.217070 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.217367 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.717356212 +0000 UTC m=+146.698595664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.318317 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.318542 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.818513392 +0000 UTC m=+146.799752844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.318788 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.319082 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.819070488 +0000 UTC m=+146.800309930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.386260 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" event={"ID":"a35c3bf5-d486-4502-9a31-0dc7848e4bd8","Type":"ContainerStarted","Data":"acd54857c52ecbf65871d2af32d391f4bfa95d90a8de8345fd79855fab71f0a2"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.388543 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" event={"ID":"42350c04-b619-45e9-8567-8f71d4c95e46","Type":"ContainerStarted","Data":"384109ca600939548edf6bd1b88eca114b207be9fba608f82d3e72f6666de7f6"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.388586 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" event={"ID":"42350c04-b619-45e9-8567-8f71d4c95e46","Type":"ContainerStarted","Data":"df90f48f77862de1486abe9dc0e863ef91664a4a29e9d537009f284df6183f10"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.390344 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" event={"ID":"5b689e3e-c89a-42bc-b6df-180d587cc197","Type":"ContainerStarted","Data":"f389635ee09a4e286a8841af71b84f7decf6df67eefe901138c4fe190e80a694"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.392131 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" event={"ID":"99337d25-bf3b-4740-9e30-8ef50cecd6cd","Type":"ContainerStarted","Data":"121b9d6dc4a7a2dbc26cb89e1ca7a2ca653c645f1570d14f5f058cde3f3ce973"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.392458 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.394061 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" event={"ID":"9aa1836d-6217-4c0f-bb26-dfc40643305d","Type":"ContainerStarted","Data":"50cbf965c3895a16d57a5a49d74836fd80c6e615fb1290a2f4c99c672b475e35"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.406477 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" event={"ID":"059e5978-1a83-45f0-876a-c497fa2365b6","Type":"ContainerStarted","Data":"bd43f05d918ebd153e886e4e1c9772bbbc965a363c1d96b7dfb3c3d55c955fd7"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.406522 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" event={"ID":"059e5978-1a83-45f0-876a-c497fa2365b6","Type":"ContainerStarted","Data":"8445b8e2fb9a3d18e41deb6497b5f768b6bdae03f12db4e9c064b33dd54987c6"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.415885 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" event={"ID":"a51dd00a-6c0c-4a57-a1f9-cbfd0d15eb84","Type":"ContainerStarted","Data":"cd228ecb4e02e672926858af6863587d701c71d9edc9fbd948c6c999922e1cff"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.417871 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" event={"ID":"89e82ba1-52a9-4c22-bc43-7d71bf835d2f","Type":"ContainerStarted","Data":"6deb542e30ef6bb578abb317600aa20acbfcbe569760ca1b486f2328d47f8f8c"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.417913 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" event={"ID":"89e82ba1-52a9-4c22-bc43-7d71bf835d2f","Type":"ContainerStarted","Data":"6f34640d8f749aa93a4010cc3e3b303ec8f9c15509972ece0ffbe20d8de09fc9"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.419264 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" podStartSLOduration=120.41924866 podStartE2EDuration="2m0.41924866s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:32.56399681 +0000 UTC m=+145.545236272" watchObservedRunningTime="2025-10-14 07:53:33.41924866 +0000 UTC m=+146.400488122" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.419472 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.420229 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:33.920205597 +0000 UTC m=+146.901445089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.420291 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-tk88f" podStartSLOduration=119.420285229 podStartE2EDuration="1m59.420285229s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.417764789 +0000 UTC m=+146.399004241" watchObservedRunningTime="2025-10-14 07:53:33.420285229 +0000 UTC m=+146.401524671" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.426197 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-l8k2s" event={"ID":"b6350740-ff4e-484d-b816-848f96e98525","Type":"ContainerStarted","Data":"934404cbcfad107013448af11c79ab35658a7754c5af60f69cebe6f966ec7cc2"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.431944 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" event={"ID":"f6a3e3bd-467a-4523-89d3-0fb063ba2a94","Type":"ContainerStarted","Data":"e13dbac91693c2986df56a6cde2695f4d63c696fe9d46985e1861602e7106bd9"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.432182 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.433466 5002 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5g28n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.433509 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" podUID="f6a3e3bd-467a-4523-89d3-0fb063ba2a94" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.433568 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" event={"ID":"3fdd526f-086e-4fb1-8a38-d297ae1843e6","Type":"ContainerStarted","Data":"f94b53301c85a54352c7af0d5c894ed55c2f6516dc87bca9a21264d5d104ba30"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.435414 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" event={"ID":"dd63d35e-85fa-44ae-9389-28b816cffae3","Type":"ContainerStarted","Data":"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.435890 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.438027 5002 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ndpjk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.438070 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.438518 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" event={"ID":"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395","Type":"ContainerStarted","Data":"5df123e696c9693e8ba06e7aa412520c723cb4e50ab9498cbcc4bc0f74a6c86d"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.438544 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" event={"ID":"1a0b16d3-a2ae-4c97-a0cf-e3ba76822395","Type":"ContainerStarted","Data":"9d4838075dc097f8a68258319d44ee9bfc90f312053b9930646db87373df0af4"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.447086 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" event={"ID":"12829eb0-c7da-49ac-8350-ce7329fded6b","Type":"ContainerStarted","Data":"fd1c161570f95cfb96634c834d3fe98dd93ffdd74cd96e50a253fa6a281a81c7"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.448433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" event={"ID":"8f63d16a-af72-4d47-828d-8ac2532a6196","Type":"ContainerStarted","Data":"7c39131f69a4475e22d2bbdfba61f63c57ae6c93d2ec3e1ff522c9724b63aa2d"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.449818 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" event={"ID":"ce76625d-d545-4664-b365-97cd648cb7ab","Type":"ContainerStarted","Data":"5d09fae121d854264f047589d76973846b75d98fc5df19fd85a27da0d43f47da"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.451781 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" event={"ID":"de309abf-bc67-416d-b0b0-1da8f0e952a2","Type":"ContainerStarted","Data":"a9ce9beeb1082b99cbeebb8ecba2368c4cd338d3c9727c1b8f5602868425daad"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.454622 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-s6s5j" event={"ID":"74ba69e9-c389-416e-aef3-9645524e6944","Type":"ContainerStarted","Data":"7bbc4a505512e0397839aebfe481a303e28da6bf6777426d56055442d235f94a"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.467999 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rjkc7" event={"ID":"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a","Type":"ContainerStarted","Data":"6a37112886f3b2cce9384b7df8f6d465d253c75b8e31035898004c06356dfe2b"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.468038 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rjkc7" event={"ID":"2d0b4cdd-b5c1-46ad-a1db-1780877eb35a","Type":"ContainerStarted","Data":"101c41b3d1e74c6d0d1ad0a755b7dbab33b58c742c486984d4ccfe632c6a0d7c"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.468540 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.471074 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xdlrk" event={"ID":"3ccad502-97aa-44b3-bb4f-d981ec035d20","Type":"ContainerStarted","Data":"49f940ccc5aeefa9be17778c69f2aefc96251bfe526dcce2e318abcdf6d7f7cb"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.472408 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.474948 5002 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdlrk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.474999 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdlrk" podUID="3ccad502-97aa-44b3-bb4f-d981ec035d20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.475222 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:33 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:33 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:33 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.475396 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.488988 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" event={"ID":"5b600621-f833-48ef-9d9d-04f6543d81ae","Type":"ContainerStarted","Data":"6fbbe0529394993814331ab88a467a4f74c08ad4ba075f09a55e76243fa79831"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.489213 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" event={"ID":"5b600621-f833-48ef-9d9d-04f6543d81ae","Type":"ContainerStarted","Data":"72617f102f5680da1af92bf85cf04be6f523f808c982275c413203815f335ad6"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.496387 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zjgv5" podStartSLOduration=119.49637293 podStartE2EDuration="1m59.49637293s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.495293691 +0000 UTC m=+146.476533143" watchObservedRunningTime="2025-10-14 07:53:33.49637293 +0000 UTC m=+146.477612382" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.497600 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" podStartSLOduration=119.497592824 podStartE2EDuration="1m59.497592824s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.459514923 +0000 UTC m=+146.440754375" watchObservedRunningTime="2025-10-14 07:53:33.497592824 +0000 UTC m=+146.478832286" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.502000 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" event={"ID":"4f691612-f354-4962-875d-8ff20478a22d","Type":"ContainerStarted","Data":"61a14c979006a1357cf827e5a3e82cc14cf7eac87108da196c4ecc886e5aca3e"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.515700 5002 generic.go:334] "Generic (PLEG): container finished" podID="20a25551-3a54-4eb2-8e5a-e13134f2ea95" containerID="9470f548a3aee49349a4b058bf82ee0849782e5a67300879407d8423a5765c43" exitCode=0 Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.515955 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" event={"ID":"20a25551-3a54-4eb2-8e5a-e13134f2ea95","Type":"ContainerDied","Data":"9470f548a3aee49349a4b058bf82ee0849782e5a67300879407d8423a5765c43"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.522550 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.528746 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.028731693 +0000 UTC m=+147.009971145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.530651 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" event={"ID":"2a8570c9-12e7-488e-aeac-c6c566122868","Type":"ContainerStarted","Data":"6002d529e169a72bc132b9576406b9bd367e0199f62b52391d89afb082e229ea"} Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.532209 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.544072 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-l9m77" podStartSLOduration=119.544051689 podStartE2EDuration="1m59.544051689s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.539431161 +0000 UTC m=+146.520670613" watchObservedRunningTime="2025-10-14 07:53:33.544051689 +0000 UTC m=+146.525291141" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.575150 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dg9pb" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.605622 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" podStartSLOduration=120.605604455 podStartE2EDuration="2m0.605604455s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.592275354 +0000 UTC m=+146.573514816" watchObservedRunningTime="2025-10-14 07:53:33.605604455 +0000 UTC m=+146.586843907" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.614211 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dvnpb" podStartSLOduration=119.614195865 podStartE2EDuration="1m59.614195865s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.613564077 +0000 UTC m=+146.594803529" watchObservedRunningTime="2025-10-14 07:53:33.614195865 +0000 UTC m=+146.595435317" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.624358 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.626544 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.126525028 +0000 UTC m=+147.107764480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.685677 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-tf7xq" podStartSLOduration=119.685658518 podStartE2EDuration="1m59.685658518s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.659265892 +0000 UTC m=+146.640505354" watchObservedRunningTime="2025-10-14 07:53:33.685658518 +0000 UTC m=+146.666897970" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.717045 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xdlrk" podStartSLOduration=120.717026281 podStartE2EDuration="2m0.717026281s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.715771986 +0000 UTC m=+146.697011448" watchObservedRunningTime="2025-10-14 07:53:33.717026281 +0000 UTC m=+146.698265733" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.725884 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.726431 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.226420644 +0000 UTC m=+147.207660096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.747718 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jpw66" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.776690 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" podStartSLOduration=119.776675775 podStartE2EDuration="1m59.776675775s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.775188113 +0000 UTC m=+146.756427575" watchObservedRunningTime="2025-10-14 07:53:33.776675775 +0000 UTC m=+146.757915227" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.777326 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-l8k2s" podStartSLOduration=7.777319642 podStartE2EDuration="7.777319642s" podCreationTimestamp="2025-10-14 07:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.748082297 +0000 UTC m=+146.729321759" watchObservedRunningTime="2025-10-14 07:53:33.777319642 +0000 UTC m=+146.758559094" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.816141 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.827879 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.828261 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.328246952 +0000 UTC m=+147.309486404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.841304 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rjkc7" podStartSLOduration=6.841290746 podStartE2EDuration="6.841290746s" podCreationTimestamp="2025-10-14 07:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.810822726 +0000 UTC m=+146.792062168" watchObservedRunningTime="2025-10-14 07:53:33.841290746 +0000 UTC m=+146.822530198" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.859895 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" podStartSLOduration=120.859876914 podStartE2EDuration="2m0.859876914s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.856610793 +0000 UTC m=+146.837850245" watchObservedRunningTime="2025-10-14 07:53:33.859876914 +0000 UTC m=+146.841116366" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.902340 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sdc6q" podStartSLOduration=119.902322547 podStartE2EDuration="1m59.902322547s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.901665129 +0000 UTC m=+146.882904591" watchObservedRunningTime="2025-10-14 07:53:33.902322547 +0000 UTC m=+146.883561999" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.902824 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6pz4b" podStartSLOduration=119.902817511 podStartE2EDuration="1m59.902817511s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.871084646 +0000 UTC m=+146.852324098" watchObservedRunningTime="2025-10-14 07:53:33.902817511 +0000 UTC m=+146.884056973" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.905405 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.905566 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.929540 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:33 crc kubenswrapper[5002]: E1014 07:53:33.929946 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.429933137 +0000 UTC m=+147.411172589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.951766 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9httn" podStartSLOduration=119.951752075 podStartE2EDuration="1m59.951752075s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.951280471 +0000 UTC m=+146.932519943" watchObservedRunningTime="2025-10-14 07:53:33.951752075 +0000 UTC m=+146.932991527" Oct 14 07:53:33 crc kubenswrapper[5002]: I1014 07:53:33.989424 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gwzw5" podStartSLOduration=119.989406745 podStartE2EDuration="1m59.989406745s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:33.98782454 +0000 UTC m=+146.969064002" watchObservedRunningTime="2025-10-14 07:53:33.989406745 +0000 UTC m=+146.970646197" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.033330 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.033703 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.533687199 +0000 UTC m=+147.514926651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.045164 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.045423 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.052104 5002 patch_prober.go:28] interesting pod/apiserver-76f77b778f-pd9r6 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.052147 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" podUID="42350c04-b619-45e9-8567-8f71d4c95e46" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.135097 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.135442 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.635431145 +0000 UTC m=+147.616670597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.137786 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6wv7p" podStartSLOduration=120.137770101 podStartE2EDuration="2m0.137770101s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.135649641 +0000 UTC m=+147.116889103" watchObservedRunningTime="2025-10-14 07:53:34.137770101 +0000 UTC m=+147.119009553" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.188589 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4r89t" podStartSLOduration=120.188574707 podStartE2EDuration="2m0.188574707s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.184228096 +0000 UTC m=+147.165467668" watchObservedRunningTime="2025-10-14 07:53:34.188574707 +0000 UTC m=+147.169814159" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.236142 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.236322 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.736297067 +0000 UTC m=+147.717536519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.236435 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.236863 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.736856062 +0000 UTC m=+147.718095514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.277372 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" podStartSLOduration=120.277354481 podStartE2EDuration="2m0.277354481s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.27337444 +0000 UTC m=+147.254613902" watchObservedRunningTime="2025-10-14 07:53:34.277354481 +0000 UTC m=+147.258593933" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.278122 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5s65x" podStartSLOduration=120.278116113 podStartE2EDuration="2m0.278116113s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.223227993 +0000 UTC m=+147.204467445" watchObservedRunningTime="2025-10-14 07:53:34.278116113 +0000 UTC m=+147.259355565" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.308595 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fv4sf" podStartSLOduration=120.308580752 podStartE2EDuration="2m0.308580752s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.306178215 +0000 UTC m=+147.287417667" watchObservedRunningTime="2025-10-14 07:53:34.308580752 +0000 UTC m=+147.289820204" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.337214 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" podStartSLOduration=120.33719848 podStartE2EDuration="2m0.33719848s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.335580955 +0000 UTC m=+147.316820417" watchObservedRunningTime="2025-10-14 07:53:34.33719848 +0000 UTC m=+147.318437932" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.338144 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.338534 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.838505866 +0000 UTC m=+147.819745318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.338621 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.339007 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.83899959 +0000 UTC m=+147.820239042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.365585 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-s6s5j" podStartSLOduration=8.36556749 podStartE2EDuration="8.36556749s" podCreationTimestamp="2025-10-14 07:53:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.36088978 +0000 UTC m=+147.342129242" watchObservedRunningTime="2025-10-14 07:53:34.36556749 +0000 UTC m=+147.346806942" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.440434 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.440830 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.940794898 +0000 UTC m=+147.922034350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.440897 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.441403 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:34.941387284 +0000 UTC m=+147.922626726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.464696 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.472012 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:34 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:34 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:34 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.472080 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.542670 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.542853 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.042815592 +0000 UTC m=+148.024055034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.543141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.543439 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.043430249 +0000 UTC m=+148.024669691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.543541 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" event={"ID":"3566d84f-2e3b-4f11-ad84-6f33541eced7","Type":"ContainerStarted","Data":"7a5b78584bcd446676f9c375cff64cd0360baf62421c6c901dc8fabc5b7bc767"} Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.543623 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" event={"ID":"3566d84f-2e3b-4f11-ad84-6f33541eced7","Type":"ContainerStarted","Data":"ecaee29a63225ecfbb83550a2793192f39fa842354eca5431c058bac60d79a68"} Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.545331 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" event={"ID":"20a25551-3a54-4eb2-8e5a-e13134f2ea95","Type":"ContainerStarted","Data":"c2c0ce9e2507a2ad4a0562cb3ef8ebcd92f58b9c58683bf2210ab3317c27fa7a"} Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.546065 5002 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ndpjk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.546117 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.549102 5002 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdlrk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.549150 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdlrk" podUID="3ccad502-97aa-44b3-bb4f-d981ec035d20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.551518 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.565031 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h4xjg" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.579048 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" podStartSLOduration=121.579006181 podStartE2EDuration="2m1.579006181s" podCreationTimestamp="2025-10-14 07:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:34.576910742 +0000 UTC m=+147.558150194" watchObservedRunningTime="2025-10-14 07:53:34.579006181 +0000 UTC m=+147.560245643" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.646275 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.648579 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.148564909 +0000 UTC m=+148.129804361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.650393 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.652163 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5g28n" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.652950 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.152939241 +0000 UTC m=+148.134178693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.752359 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.752898 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.252883388 +0000 UTC m=+148.234122840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.854158 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.854254 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.854946 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.354929152 +0000 UTC m=+148.336168604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.872334 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.955664 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.955817 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.955961 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:34 crc kubenswrapper[5002]: E1014 07:53:34.956778 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.45676109 +0000 UTC m=+148.438000542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.957019 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:34 crc kubenswrapper[5002]: I1014 07:53:34.960706 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.056876 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.057224 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.058387 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.558371043 +0000 UTC m=+148.539610495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.063510 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.145368 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.158317 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.158509 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.658476595 +0000 UTC m=+148.639716047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.161169 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.173967 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.247571 5002 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.262526 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.262806 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.762787522 +0000 UTC m=+148.744026964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.363152 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.363456 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.863441468 +0000 UTC m=+148.844680920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.464111 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.464444 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:35.964432863 +0000 UTC m=+148.945672315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.479404 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:35 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:35 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:35 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.479451 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.561291 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b5f3bf2d874a1413eb2136cdc3d7812d9651ea818e2b64f51ff1035011da833c"} Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.564918 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.565366 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:36.065350736 +0000 UTC m=+149.046590188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.567141 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" event={"ID":"3566d84f-2e3b-4f11-ad84-6f33541eced7","Type":"ContainerStarted","Data":"4f7cd5d9a5aa7e2d192105de6710fa7eeddf626090d25b49d996e9b363a9767d"} Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.567167 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" event={"ID":"3566d84f-2e3b-4f11-ad84-6f33541eced7","Type":"ContainerStarted","Data":"d5752292112eaf79ba24846d528c1845c4120d1d4e81015f8ff74b65a59590a4"} Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.569490 5002 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdlrk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.569546 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdlrk" podUID="3ccad502-97aa-44b3-bb4f-d981ec035d20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.598390 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.613641 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cnwsm" podStartSLOduration=8.613625842 podStartE2EDuration="8.613625842s" podCreationTimestamp="2025-10-14 07:53:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:35.608238562 +0000 UTC m=+148.589478014" watchObservedRunningTime="2025-10-14 07:53:35.613625842 +0000 UTC m=+148.594865294" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.676562 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.680363 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:36.180348832 +0000 UTC m=+149.161588284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.780459 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.780783 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:53:36.280762411 +0000 UTC m=+149.262001863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: W1014 07:53:35.816948 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b4a7db1d9cf3878c087c39ee5e9481be422984af803bf14612a4386ce90aaa1e WatchSource:0}: Error finding container b4a7db1d9cf3878c087c39ee5e9481be422984af803bf14612a4386ce90aaa1e: Status 404 returned error can't find the container with id b4a7db1d9cf3878c087c39ee5e9481be422984af803bf14612a4386ce90aaa1e Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.882293 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:35 crc kubenswrapper[5002]: E1014 07:53:35.882632 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:53:36.382617551 +0000 UTC m=+149.363857003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dc64m" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.895591 5002 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-14T07:53:35.247595898Z","Handler":null,"Name":""} Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.898539 5002 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.898659 5002 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.935268 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.938536 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.941967 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.943897 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.983020 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.983190 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.983259 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc677\" (UniqueName: \"kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:35 crc kubenswrapper[5002]: I1014 07:53:35.983319 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084088 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084139 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084183 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc677\" (UniqueName: \"kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084212 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084720 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.084740 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.105943 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc677\" (UniqueName: \"kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677\") pod \"certified-operators-9xr4f\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.128805 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.129688 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.132652 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.144554 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.185366 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.185465 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk28k\" (UniqueName: \"kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.185495 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.185524 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.189584 5002 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.189610 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.210736 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dc64m\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.256109 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.286782 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.286888 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.286974 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk28k\" (UniqueName: \"kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.287440 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.287492 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.303233 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk28k\" (UniqueName: \"kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k\") pod \"community-operators-cqzkm\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.306218 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.329870 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.333368 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.347930 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.387867 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9bgj\" (UniqueName: \"kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.387909 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.387924 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.438588 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.442344 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.474136 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:36 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:36 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:36 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.474184 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.489020 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9bgj\" (UniqueName: \"kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.489054 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.489140 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.490045 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.490617 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.517393 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9bgj\" (UniqueName: \"kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj\") pod \"certified-operators-84p8s\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.518023 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.527357 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.528296 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.536909 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:53:36 crc kubenswrapper[5002]: W1014 07:53:36.608569 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53e3879b_b0b6_4bfa_876d_765526e2e952.slice/crio-ab339dabd582727f379d5add539f7c022738a03784e5154c9ddfbd5d3e3ef8c4 WatchSource:0}: Error finding container ab339dabd582727f379d5add539f7c022738a03784e5154c9ddfbd5d3e3ef8c4: Status 404 returned error can't find the container with id ab339dabd582727f379d5add539f7c022738a03784e5154c9ddfbd5d3e3ef8c4 Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.617456 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"050539dadb7d2e92bb288b7f88574a4d42ff1534a297e88896d432069f84e0c0"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.620124 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bcf8946f13c7ef759c2e6330b90037db388bd7be638c320d72e3362490102e50"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.620167 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b4a7db1d9cf3878c087c39ee5e9481be422984af803bf14612a4386ce90aaa1e"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.621745 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerStarted","Data":"e3549941f7be81c945457badd004a5de1d2036800041bec6386e9f2aa03a92ce"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.624095 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"005991c06f80da820d8b835d229249c10cc502aba220759f50c4a90b226760fb"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.624118 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ebcfd58c891dec9363063e0a36641ce8b33559aaba91767ddbad4d105e6caab3"} Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.624413 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.640342 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-s656d" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.654019 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.702606 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8mkl\" (UniqueName: \"kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.702652 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.702719 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.804471 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.806434 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.806551 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8mkl\" (UniqueName: \"kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.806593 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.808895 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.809730 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.826640 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8mkl\" (UniqueName: \"kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl\") pod \"community-operators-dbrb2\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.925427 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:53:36 crc kubenswrapper[5002]: W1014 07:53:36.936852 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3494b1ed_4ee0_4c88_95b7_183d4ed79660.slice/crio-5a50c6af2a7a9f9cefcddd9e62c20baa47906de86d42cac8037c1bc27f7c3025 WatchSource:0}: Error finding container 5a50c6af2a7a9f9cefcddd9e62c20baa47906de86d42cac8037c1bc27f7c3025: Status 404 returned error can't find the container with id 5a50c6af2a7a9f9cefcddd9e62c20baa47906de86d42cac8037c1bc27f7c3025 Oct 14 07:53:36 crc kubenswrapper[5002]: I1014 07:53:36.949191 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.132925 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:53:37 crc kubenswrapper[5002]: W1014 07:53:37.161461 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab6fbed_7856_4d82_baa7_5507198362d6.slice/crio-c043ddc4d09754551e0ff50ecd4d4d661d5da5bee3378c0d102aef9ab4cd54e1 WatchSource:0}: Error finding container c043ddc4d09754551e0ff50ecd4d4d661d5da5bee3378c0d102aef9ab4cd54e1: Status 404 returned error can't find the container with id c043ddc4d09754551e0ff50ecd4d4d661d5da5bee3378c0d102aef9ab4cd54e1 Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.473540 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:37 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:37 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:37 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.473638 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.629602 5002 generic.go:334] "Generic (PLEG): container finished" podID="eab6fbed-7856-4d82-baa7-5507198362d6" containerID="b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6" exitCode=0 Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.629702 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerDied","Data":"b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.629746 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerStarted","Data":"c043ddc4d09754551e0ff50ecd4d4d661d5da5bee3378c0d102aef9ab4cd54e1"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.630793 5002 generic.go:334] "Generic (PLEG): container finished" podID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerID="26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541" exitCode=0 Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.630864 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerDied","Data":"26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.630890 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerStarted","Data":"d3ef6124fba4ec3e1ae5fd78df81d2fbaa1c83177a948836afb0123817251d84"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.631687 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.633471 5002 generic.go:334] "Generic (PLEG): container finished" podID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerID="33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8" exitCode=0 Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.633575 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerDied","Data":"33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.637903 5002 generic.go:334] "Generic (PLEG): container finished" podID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerID="b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4" exitCode=0 Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.638005 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerDied","Data":"b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.638036 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerStarted","Data":"5a50c6af2a7a9f9cefcddd9e62c20baa47906de86d42cac8037c1bc27f7c3025"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.640140 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" event={"ID":"53e3879b-b0b6-4bfa-876d-765526e2e952","Type":"ContainerStarted","Data":"fd71c15f1884d3e4a94debda691233a4eb17220dd8a3be032426878a9d691656"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.640172 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" event={"ID":"53e3879b-b0b6-4bfa-876d-765526e2e952","Type":"ContainerStarted","Data":"ab339dabd582727f379d5add539f7c022738a03784e5154c9ddfbd5d3e3ef8c4"} Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.640496 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.734982 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.756909 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.757693 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.759274 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.759385 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.771264 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.777452 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" podStartSLOduration=123.777435811 podStartE2EDuration="2m3.777435811s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:53:37.777075442 +0000 UTC m=+150.758314924" watchObservedRunningTime="2025-10-14 07:53:37.777435811 +0000 UTC m=+150.758675273" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.923238 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:37 crc kubenswrapper[5002]: I1014 07:53:37.923280 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.024114 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.024157 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.024224 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.043568 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.072110 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.129138 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.130232 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.135414 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.139727 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.227300 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.227381 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.227414 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg96l\" (UniqueName: \"kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.320538 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.328362 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.328424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.328464 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg96l\" (UniqueName: \"kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.329154 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.329316 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.369307 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg96l\" (UniqueName: \"kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l\") pod \"redhat-marketplace-jlfz8\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.452866 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.472874 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:38 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:38 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:38 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.473196 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.526162 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.529645 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.542051 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.631217 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5hhd\" (UniqueName: \"kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.631322 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.631356 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.667551 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a5cf06f4-e41a-4b99-8631-65f0d94496fb","Type":"ContainerStarted","Data":"feb3feb7a86175622b763fc1360326caa9a2ca4c63748c47f1ea08dae31dca2b"} Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.732080 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5hhd\" (UniqueName: \"kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.732395 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.732424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.732960 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.732999 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.768720 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5hhd\" (UniqueName: \"kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd\") pod \"redhat-marketplace-fq5jf\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:38 crc kubenswrapper[5002]: E1014 07:53:38.833143 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:53:38 crc kubenswrapper[5002]: I1014 07:53:38.852471 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.031918 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.049731 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.062181 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pd9r6" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.134542 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.145967 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.146119 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.150369 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.162772 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.206727 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.206763 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:39 crc kubenswrapper[5002]: W1014 07:53:39.212859 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf427bfd_0d50_43b6_9d26_aa11ee3b149a.slice/crio-67a2b3d1354f9aa011f6781848e3eb729c0bbba1f7d19c981ff9d18d40c8ee1e WatchSource:0}: Error finding container 67a2b3d1354f9aa011f6781848e3eb729c0bbba1f7d19c981ff9d18d40c8ee1e: Status 404 returned error can't find the container with id 67a2b3d1354f9aa011f6781848e3eb729c0bbba1f7d19c981ff9d18d40c8ee1e Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.215336 5002 patch_prober.go:28] interesting pod/console-f9d7485db-sn4pd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.215391 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sn4pd" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.217758 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.217776 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.244589 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxvp4\" (UniqueName: \"kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.244999 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.245550 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.346694 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxvp4\" (UniqueName: \"kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.346748 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.346856 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.347724 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.347811 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.373470 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxvp4\" (UniqueName: \"kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4\") pod \"redhat-operators-pgl6v\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.470001 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.475378 5002 patch_prober.go:28] interesting pod/router-default-5444994796-dndgk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:53:39 crc kubenswrapper[5002]: [-]has-synced failed: reason withheld Oct 14 07:53:39 crc kubenswrapper[5002]: [+]process-running ok Oct 14 07:53:39 crc kubenswrapper[5002]: healthz check failed Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.475434 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dndgk" podUID="9f6b37d8-9ab1-41e6-a511-3116e9441ab2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.516084 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.531128 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.532110 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.541630 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.656866 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.656913 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.656975 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5ns9\" (UniqueName: \"kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.714781 5002 generic.go:334] "Generic (PLEG): container finished" podID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerID="7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610" exitCode=0 Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.714988 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerDied","Data":"7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610"} Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.715041 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerStarted","Data":"02fa4e0cba546e1b1e1f7b344538fcfc29712a714b58aae8c6113b77978737a3"} Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.733763 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerID="70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459" exitCode=0 Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.736227 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerDied","Data":"70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459"} Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.736270 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerStarted","Data":"67a2b3d1354f9aa011f6781848e3eb729c0bbba1f7d19c981ff9d18d40c8ee1e"} Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.744785 5002 generic.go:334] "Generic (PLEG): container finished" podID="a5cf06f4-e41a-4b99-8631-65f0d94496fb" containerID="0329d1359278e78f02b0f0cbccbb05635f22424a4cc760ca69381ce6755343fc" exitCode=0 Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.746976 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a5cf06f4-e41a-4b99-8631-65f0d94496fb","Type":"ContainerDied","Data":"0329d1359278e78f02b0f0cbccbb05635f22424a4cc760ca69381ce6755343fc"} Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.757664 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.757733 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5ns9\" (UniqueName: \"kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.757802 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.758259 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.758454 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.776054 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5ns9\" (UniqueName: \"kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9\") pod \"redhat-operators-dq285\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:39 crc kubenswrapper[5002]: I1014 07:53:39.850883 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.110166 5002 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdlrk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.110494 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdlrk" podUID="3ccad502-97aa-44b3-bb4f-d981ec035d20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.110397 5002 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdlrk container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.110940 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xdlrk" podUID="3ccad502-97aa-44b3-bb4f-d981ec035d20" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.42:8080/\": dial tcp 10.217.0.42:8080: connect: connection refused" Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.162466 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.171161 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:53:40 crc kubenswrapper[5002]: W1014 07:53:40.185571 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd40e9f14_1956_43e3_a22c_4fcc1e5f6450.slice/crio-c7031efc41bc111e428e70737cc10a3784eeb9b614253199d695c06956d7c15b WatchSource:0}: Error finding container c7031efc41bc111e428e70737cc10a3784eeb9b614253199d695c06956d7c15b: Status 404 returned error can't find the container with id c7031efc41bc111e428e70737cc10a3784eeb9b614253199d695c06956d7c15b Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.484448 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.493186 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-dndgk" Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.791624 5002 generic.go:334] "Generic (PLEG): container finished" podID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerID="2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540" exitCode=0 Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.791696 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerDied","Data":"2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540"} Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.791720 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerStarted","Data":"c7031efc41bc111e428e70737cc10a3784eeb9b614253199d695c06956d7c15b"} Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.812543 5002 generic.go:334] "Generic (PLEG): container finished" podID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerID="8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a" exitCode=0 Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.813375 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerDied","Data":"8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a"} Oct 14 07:53:40 crc kubenswrapper[5002]: I1014 07:53:40.813396 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerStarted","Data":"cb58630f27289c8c8bc46c28faf72cfbf33e5ab8710544142474e5a000258ade"} Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.242758 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.325964 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:53:41 crc kubenswrapper[5002]: E1014 07:53:41.326348 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cf06f4-e41a-4b99-8631-65f0d94496fb" containerName="pruner" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.326370 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cf06f4-e41a-4b99-8631-65f0d94496fb" containerName="pruner" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.326477 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cf06f4-e41a-4b99-8631-65f0d94496fb" containerName="pruner" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.326872 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.326945 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.331223 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.331285 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397098 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir\") pod \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397181 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access\") pod \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\" (UID: \"a5cf06f4-e41a-4b99-8631-65f0d94496fb\") " Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397254 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a5cf06f4-e41a-4b99-8631-65f0d94496fb" (UID: "a5cf06f4-e41a-4b99-8631-65f0d94496fb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397711 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397739 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.397897 5002 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.416013 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a5cf06f4-e41a-4b99-8631-65f0d94496fb" (UID: "a5cf06f4-e41a-4b99-8631-65f0d94496fb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.499903 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.499938 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.500368 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5cf06f4-e41a-4b99-8631-65f0d94496fb-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.500469 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.518723 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.653933 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.821959 5002 generic.go:334] "Generic (PLEG): container finished" podID="8f63d16a-af72-4d47-828d-8ac2532a6196" containerID="7c39131f69a4475e22d2bbdfba61f63c57ae6c93d2ec3e1ff522c9724b63aa2d" exitCode=0 Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.822025 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" event={"ID":"8f63d16a-af72-4d47-828d-8ac2532a6196","Type":"ContainerDied","Data":"7c39131f69a4475e22d2bbdfba61f63c57ae6c93d2ec3e1ff522c9724b63aa2d"} Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.828739 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a5cf06f4-e41a-4b99-8631-65f0d94496fb","Type":"ContainerDied","Data":"feb3feb7a86175622b763fc1360326caa9a2ca4c63748c47f1ea08dae31dca2b"} Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.828787 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feb3feb7a86175622b763fc1360326caa9a2ca4c63748c47f1ea08dae31dca2b" Oct 14 07:53:41 crc kubenswrapper[5002]: I1014 07:53:41.828867 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:53:42 crc kubenswrapper[5002]: I1014 07:53:42.084729 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:53:42 crc kubenswrapper[5002]: W1014 07:53:42.124424 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb09b338a_19a0_4b68_9774_aeb15f68e63c.slice/crio-5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141 WatchSource:0}: Error finding container 5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141: Status 404 returned error can't find the container with id 5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141 Oct 14 07:53:42 crc kubenswrapper[5002]: I1014 07:53:42.156092 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rjkc7" Oct 14 07:53:42 crc kubenswrapper[5002]: I1014 07:53:42.837893 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b09b338a-19a0-4b68-9774-aeb15f68e63c","Type":"ContainerStarted","Data":"5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141"} Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.148186 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.326139 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb4gq\" (UniqueName: \"kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq\") pod \"8f63d16a-af72-4d47-828d-8ac2532a6196\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.326201 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume\") pod \"8f63d16a-af72-4d47-828d-8ac2532a6196\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.326235 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume\") pod \"8f63d16a-af72-4d47-828d-8ac2532a6196\" (UID: \"8f63d16a-af72-4d47-828d-8ac2532a6196\") " Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.329219 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume" (OuterVolumeSpecName: "config-volume") pod "8f63d16a-af72-4d47-828d-8ac2532a6196" (UID: "8f63d16a-af72-4d47-828d-8ac2532a6196"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.333196 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq" (OuterVolumeSpecName: "kube-api-access-rb4gq") pod "8f63d16a-af72-4d47-828d-8ac2532a6196" (UID: "8f63d16a-af72-4d47-828d-8ac2532a6196"). InnerVolumeSpecName "kube-api-access-rb4gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.339320 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8f63d16a-af72-4d47-828d-8ac2532a6196" (UID: "8f63d16a-af72-4d47-828d-8ac2532a6196"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.427850 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb4gq\" (UniqueName: \"kubernetes.io/projected/8f63d16a-af72-4d47-828d-8ac2532a6196-kube-api-access-rb4gq\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.427879 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8f63d16a-af72-4d47-828d-8ac2532a6196-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.427890 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8f63d16a-af72-4d47-828d-8ac2532a6196-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.856591 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" event={"ID":"8f63d16a-af72-4d47-828d-8ac2532a6196","Type":"ContainerDied","Data":"b9a0996b422906051ce44467ea563c302798b7cc5323f81db3bdefc5102bbe23"} Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.856629 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9a0996b422906051ce44467ea563c302798b7cc5323f81db3bdefc5102bbe23" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.856688 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm" Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.874265 5002 generic.go:334] "Generic (PLEG): container finished" podID="b09b338a-19a0-4b68-9774-aeb15f68e63c" containerID="e5bec8558b515eb163ea177177ac445d5fd805dd4b262bd9ed867bad71e2c603" exitCode=0 Oct 14 07:53:43 crc kubenswrapper[5002]: I1014 07:53:43.874318 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b09b338a-19a0-4b68-9774-aeb15f68e63c","Type":"ContainerDied","Data":"e5bec8558b515eb163ea177177ac445d5fd805dd4b262bd9ed867bad71e2c603"} Oct 14 07:53:46 crc kubenswrapper[5002]: I1014 07:53:46.895508 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-b7sbl_0b5fc9a0-1cd3-479a-b0b7-6a9e907128be/cluster-samples-operator/0.log" Oct 14 07:53:46 crc kubenswrapper[5002]: I1014 07:53:46.895898 5002 generic.go:334] "Generic (PLEG): container finished" podID="0b5fc9a0-1cd3-479a-b0b7-6a9e907128be" containerID="6ea6ab50857558e881c4599871deb7779e3b43075a035e8cf20ae202f7d6f20d" exitCode=2 Oct 14 07:53:46 crc kubenswrapper[5002]: I1014 07:53:46.895936 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" event={"ID":"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be","Type":"ContainerDied","Data":"6ea6ab50857558e881c4599871deb7779e3b43075a035e8cf20ae202f7d6f20d"} Oct 14 07:53:46 crc kubenswrapper[5002]: I1014 07:53:46.897226 5002 scope.go:117] "RemoveContainer" containerID="6ea6ab50857558e881c4599871deb7779e3b43075a035e8cf20ae202f7d6f20d" Oct 14 07:53:48 crc kubenswrapper[5002]: E1014 07:53:48.929500 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:53:49 crc kubenswrapper[5002]: I1014 07:53:49.207291 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:49 crc kubenswrapper[5002]: I1014 07:53:49.210386 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 07:53:50 crc kubenswrapper[5002]: I1014 07:53:50.131478 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xdlrk" Oct 14 07:53:50 crc kubenswrapper[5002]: I1014 07:53:50.922562 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b09b338a-19a0-4b68-9774-aeb15f68e63c","Type":"ContainerDied","Data":"5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141"} Oct 14 07:53:50 crc kubenswrapper[5002]: I1014 07:53:50.922639 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5208871fa7515a86be306ec7c0f981205f9e1f595987f7fbeb46f7ca85bd7141" Oct 14 07:53:50 crc kubenswrapper[5002]: I1014 07:53:50.996421 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.055674 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access\") pod \"b09b338a-19a0-4b68-9774-aeb15f68e63c\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.055814 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir\") pod \"b09b338a-19a0-4b68-9774-aeb15f68e63c\" (UID: \"b09b338a-19a0-4b68-9774-aeb15f68e63c\") " Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.056476 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b09b338a-19a0-4b68-9774-aeb15f68e63c" (UID: "b09b338a-19a0-4b68-9774-aeb15f68e63c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.060776 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b09b338a-19a0-4b68-9774-aeb15f68e63c" (UID: "b09b338a-19a0-4b68-9774-aeb15f68e63c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.157336 5002 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b09b338a-19a0-4b68-9774-aeb15f68e63c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.157369 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b09b338a-19a0-4b68-9774-aeb15f68e63c-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:53:51 crc kubenswrapper[5002]: I1014 07:53:51.928111 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:53:56 crc kubenswrapper[5002]: I1014 07:53:56.311476 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:53:56 crc kubenswrapper[5002]: I1014 07:53:56.643673 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:56 crc kubenswrapper[5002]: I1014 07:53:56.652675 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db613c5e-6bf3-4589-aca8-f6c7da7d36f9-metrics-certs\") pod \"network-metrics-daemon-f5rhq\" (UID: \"db613c5e-6bf3-4589-aca8-f6c7da7d36f9\") " pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:56 crc kubenswrapper[5002]: I1014 07:53:56.838689 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f5rhq" Oct 14 07:53:59 crc kubenswrapper[5002]: E1014 07:53:59.044107 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359daa94_9198_48cf_bbea_a7d8cdb8f3c7.slice/crio-conmon-35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:54:03 crc kubenswrapper[5002]: I1014 07:54:03.860231 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-f5rhq"] Oct 14 07:54:03 crc kubenswrapper[5002]: W1014 07:54:03.887524 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb613c5e_6bf3_4589_aca8_f6c7da7d36f9.slice/crio-2ca934a81539e9e69b0869059cb1af061b41974a8b0d7ce4083e3c7040288be3 WatchSource:0}: Error finding container 2ca934a81539e9e69b0869059cb1af061b41974a8b0d7ce4083e3c7040288be3: Status 404 returned error can't find the container with id 2ca934a81539e9e69b0869059cb1af061b41974a8b0d7ce4083e3c7040288be3 Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.007770 5002 generic.go:334] "Generic (PLEG): container finished" podID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerID="a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a" exitCode=0 Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.007894 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerDied","Data":"a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.009711 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerStarted","Data":"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.011261 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerStarted","Data":"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.015131 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerStarted","Data":"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.018440 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerStarted","Data":"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.020614 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerStarted","Data":"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.022728 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-b7sbl_0b5fc9a0-1cd3-479a-b0b7-6a9e907128be/cluster-samples-operator/0.log" Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.022801 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-b7sbl" event={"ID":"0b5fc9a0-1cd3-479a-b0b7-6a9e907128be","Type":"ContainerStarted","Data":"6993dd3098eeb918033a50010cc83015a679473d993073fb3b0bda353ea3ea7f"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.028377 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" event={"ID":"db613c5e-6bf3-4589-aca8-f6c7da7d36f9","Type":"ContainerStarted","Data":"2ca934a81539e9e69b0869059cb1af061b41974a8b0d7ce4083e3c7040288be3"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.033278 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerID="02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93" exitCode=0 Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.033352 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerDied","Data":"02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93"} Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.036370 5002 generic.go:334] "Generic (PLEG): container finished" podID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerID="67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396" exitCode=0 Oct 14 07:54:04 crc kubenswrapper[5002]: I1014 07:54:04.036408 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerDied","Data":"67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.050228 5002 generic.go:334] "Generic (PLEG): container finished" podID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerID="c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171" exitCode=0 Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.050328 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerDied","Data":"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.054796 5002 generic.go:334] "Generic (PLEG): container finished" podID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerID="5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321" exitCode=0 Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.054902 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerDied","Data":"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.060899 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" event={"ID":"db613c5e-6bf3-4589-aca8-f6c7da7d36f9","Type":"ContainerStarted","Data":"f1167099f7229a9d0b49e445f491d37fa95425d66b9feb03b4074686aa891eed"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.060980 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f5rhq" event={"ID":"db613c5e-6bf3-4589-aca8-f6c7da7d36f9","Type":"ContainerStarted","Data":"eeac29ef4a0e4bf448f62ccc58dbcc62c92fd092483c6190e6e504e78bfd672c"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.063669 5002 generic.go:334] "Generic (PLEG): container finished" podID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerID="bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67" exitCode=0 Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.063733 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerDied","Data":"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.070665 5002 generic.go:334] "Generic (PLEG): container finished" podID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerID="8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3" exitCode=0 Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.070748 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerDied","Data":"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.076159 5002 generic.go:334] "Generic (PLEG): container finished" podID="eab6fbed-7856-4d82-baa7-5507198362d6" containerID="7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a" exitCode=0 Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.076217 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerDied","Data":"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a"} Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.234354 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:54:05 crc kubenswrapper[5002]: I1014 07:54:05.255392 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-f5rhq" podStartSLOduration=151.255376167 podStartE2EDuration="2m31.255376167s" podCreationTimestamp="2025-10-14 07:51:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:54:05.221765711 +0000 UTC m=+178.203005173" watchObservedRunningTime="2025-10-14 07:54:05.255376167 +0000 UTC m=+178.236615639" Oct 14 07:54:07 crc kubenswrapper[5002]: I1014 07:54:07.093156 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerStarted","Data":"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f"} Oct 14 07:54:07 crc kubenswrapper[5002]: I1014 07:54:07.120516 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cqzkm" podStartSLOduration=2.678952613 podStartE2EDuration="31.12048032s" podCreationTimestamp="2025-10-14 07:53:36 +0000 UTC" firstStartedPulling="2025-10-14 07:53:37.632022388 +0000 UTC m=+150.613261870" lastFinishedPulling="2025-10-14 07:54:06.073550105 +0000 UTC m=+179.054789577" observedRunningTime="2025-10-14 07:54:07.119461431 +0000 UTC m=+180.100700913" watchObservedRunningTime="2025-10-14 07:54:07.12048032 +0000 UTC m=+180.101719802" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.114470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerStarted","Data":"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.117122 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerStarted","Data":"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.118696 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerStarted","Data":"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.120151 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerStarted","Data":"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.121808 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerStarted","Data":"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.135989 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerStarted","Data":"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.138942 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dbrb2" podStartSLOduration=2.209717121 podStartE2EDuration="33.138926567s" podCreationTimestamp="2025-10-14 07:53:36 +0000 UTC" firstStartedPulling="2025-10-14 07:53:37.631464632 +0000 UTC m=+150.612704084" lastFinishedPulling="2025-10-14 07:54:08.560674048 +0000 UTC m=+181.541913530" observedRunningTime="2025-10-14 07:54:09.138848095 +0000 UTC m=+182.120087547" watchObservedRunningTime="2025-10-14 07:54:09.138926567 +0000 UTC m=+182.120166019" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.145371 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerStarted","Data":"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99"} Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.169098 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pgl6v" podStartSLOduration=2.2920628450000002 podStartE2EDuration="30.169083617s" podCreationTimestamp="2025-10-14 07:53:39 +0000 UTC" firstStartedPulling="2025-10-14 07:53:40.800923615 +0000 UTC m=+153.782163067" lastFinishedPulling="2025-10-14 07:54:08.677944357 +0000 UTC m=+181.659183839" observedRunningTime="2025-10-14 07:54:09.165862858 +0000 UTC m=+182.147102330" watchObservedRunningTime="2025-10-14 07:54:09.169083617 +0000 UTC m=+182.150323069" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.186870 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jlfz8" podStartSLOduration=2.798870327 podStartE2EDuration="31.186855573s" podCreationTimestamp="2025-10-14 07:53:38 +0000 UTC" firstStartedPulling="2025-10-14 07:53:39.727358248 +0000 UTC m=+152.708597700" lastFinishedPulling="2025-10-14 07:54:08.115343454 +0000 UTC m=+181.096582946" observedRunningTime="2025-10-14 07:54:09.185036473 +0000 UTC m=+182.166275925" watchObservedRunningTime="2025-10-14 07:54:09.186855573 +0000 UTC m=+182.168095025" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.206478 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dq285" podStartSLOduration=2.438308674 podStartE2EDuration="30.20646259s" podCreationTimestamp="2025-10-14 07:53:39 +0000 UTC" firstStartedPulling="2025-10-14 07:53:40.815886952 +0000 UTC m=+153.797126404" lastFinishedPulling="2025-10-14 07:54:08.584040828 +0000 UTC m=+181.565280320" observedRunningTime="2025-10-14 07:54:09.20432979 +0000 UTC m=+182.185569272" watchObservedRunningTime="2025-10-14 07:54:09.20646259 +0000 UTC m=+182.187702042" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.217749 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.218037 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.225827 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9xr4f" podStartSLOduration=3.387523628 podStartE2EDuration="34.225810329s" podCreationTimestamp="2025-10-14 07:53:35 +0000 UTC" firstStartedPulling="2025-10-14 07:53:37.63712121 +0000 UTC m=+150.618360662" lastFinishedPulling="2025-10-14 07:54:08.475407901 +0000 UTC m=+181.456647363" observedRunningTime="2025-10-14 07:54:09.224209345 +0000 UTC m=+182.205448807" watchObservedRunningTime="2025-10-14 07:54:09.225810329 +0000 UTC m=+182.207049781" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.249991 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fq5jf" podStartSLOduration=2.544286071 podStartE2EDuration="31.249975323s" podCreationTimestamp="2025-10-14 07:53:38 +0000 UTC" firstStartedPulling="2025-10-14 07:53:39.735963948 +0000 UTC m=+152.717203400" lastFinishedPulling="2025-10-14 07:54:08.44165316 +0000 UTC m=+181.422892652" observedRunningTime="2025-10-14 07:54:09.248363148 +0000 UTC m=+182.229602610" watchObservedRunningTime="2025-10-14 07:54:09.249975323 +0000 UTC m=+182.231214775" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.271439 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84p8s" podStartSLOduration=2.350041293 podStartE2EDuration="33.27142339s" podCreationTimestamp="2025-10-14 07:53:36 +0000 UTC" firstStartedPulling="2025-10-14 07:53:37.639320271 +0000 UTC m=+150.620559723" lastFinishedPulling="2025-10-14 07:54:08.560702358 +0000 UTC m=+181.541941820" observedRunningTime="2025-10-14 07:54:09.270385122 +0000 UTC m=+182.251624574" watchObservedRunningTime="2025-10-14 07:54:09.27142339 +0000 UTC m=+182.252662842" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.433887 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5nc9n" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.516507 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.516567 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.852025 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:09 crc kubenswrapper[5002]: I1014 07:54:09.852066 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:10 crc kubenswrapper[5002]: I1014 07:54:10.797914 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pgl6v" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="registry-server" probeResult="failure" output=< Oct 14 07:54:10 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 07:54:10 crc kubenswrapper[5002]: > Oct 14 07:54:10 crc kubenswrapper[5002]: I1014 07:54:10.886569 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dq285" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="registry-server" probeResult="failure" output=< Oct 14 07:54:10 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 07:54:10 crc kubenswrapper[5002]: > Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.256971 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.258727 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.322824 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.443352 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.443417 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.488614 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.655049 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.655093 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.719072 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.950404 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:16 crc kubenswrapper[5002]: I1014 07:54:16.950441 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.024546 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.259342 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.261480 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.271791 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.275893 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:17 crc kubenswrapper[5002]: I1014 07:54:17.957370 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.453870 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.454200 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.525192 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.853636 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.854033 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:18 crc kubenswrapper[5002]: I1014 07:54:18.918999 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.208003 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-84p8s" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="registry-server" containerID="cri-o://8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99" gracePeriod=2 Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.254087 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.269919 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.356529 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.357143 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dbrb2" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="registry-server" containerID="cri-o://4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc" gracePeriod=2 Oct 14 07:54:19 crc kubenswrapper[5002]: E1014 07:54:19.361231 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3494b1ed_4ee0_4c88_95b7_183d4ed79660.slice/crio-8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3494b1ed_4ee0_4c88_95b7_183d4ed79660.slice/crio-conmon-8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.576659 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.611451 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.703727 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.773793 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content\") pod \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.774015 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9bgj\" (UniqueName: \"kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj\") pod \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.774050 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities\") pod \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\" (UID: \"3494b1ed-4ee0-4c88-95b7-183d4ed79660\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.774763 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities" (OuterVolumeSpecName: "utilities") pod "3494b1ed-4ee0-4c88-95b7-183d4ed79660" (UID: "3494b1ed-4ee0-4c88-95b7-183d4ed79660"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.780583 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj" (OuterVolumeSpecName: "kube-api-access-t9bgj") pod "3494b1ed-4ee0-4c88-95b7-183d4ed79660" (UID: "3494b1ed-4ee0-4c88-95b7-183d4ed79660"). InnerVolumeSpecName "kube-api-access-t9bgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.782142 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.816177 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3494b1ed-4ee0-4c88-95b7-183d4ed79660" (UID: "3494b1ed-4ee0-4c88-95b7-183d4ed79660"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.874945 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8mkl\" (UniqueName: \"kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl\") pod \"eab6fbed-7856-4d82-baa7-5507198362d6\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.875012 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities\") pod \"eab6fbed-7856-4d82-baa7-5507198362d6\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.875081 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content\") pod \"eab6fbed-7856-4d82-baa7-5507198362d6\" (UID: \"eab6fbed-7856-4d82-baa7-5507198362d6\") " Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.875247 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.875259 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9bgj\" (UniqueName: \"kubernetes.io/projected/3494b1ed-4ee0-4c88-95b7-183d4ed79660-kube-api-access-t9bgj\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.875270 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3494b1ed-4ee0-4c88-95b7-183d4ed79660-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.876674 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities" (OuterVolumeSpecName: "utilities") pod "eab6fbed-7856-4d82-baa7-5507198362d6" (UID: "eab6fbed-7856-4d82-baa7-5507198362d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.879303 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl" (OuterVolumeSpecName: "kube-api-access-t8mkl") pod "eab6fbed-7856-4d82-baa7-5507198362d6" (UID: "eab6fbed-7856-4d82-baa7-5507198362d6"). InnerVolumeSpecName "kube-api-access-t8mkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.899605 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.918737 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eab6fbed-7856-4d82-baa7-5507198362d6" (UID: "eab6fbed-7856-4d82-baa7-5507198362d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.945028 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.976687 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.976728 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8mkl\" (UniqueName: \"kubernetes.io/projected/eab6fbed-7856-4d82-baa7-5507198362d6-kube-api-access-t8mkl\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:19 crc kubenswrapper[5002]: I1014 07:54:19.976745 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab6fbed-7856-4d82-baa7-5507198362d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.213960 5002 generic.go:334] "Generic (PLEG): container finished" podID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerID="8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99" exitCode=0 Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.214019 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerDied","Data":"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99"} Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.214344 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84p8s" event={"ID":"3494b1ed-4ee0-4c88-95b7-183d4ed79660","Type":"ContainerDied","Data":"5a50c6af2a7a9f9cefcddd9e62c20baa47906de86d42cac8037c1bc27f7c3025"} Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.214091 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84p8s" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.214379 5002 scope.go:117] "RemoveContainer" containerID="8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.218472 5002 generic.go:334] "Generic (PLEG): container finished" podID="eab6fbed-7856-4d82-baa7-5507198362d6" containerID="4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc" exitCode=0 Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.218552 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dbrb2" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.218604 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerDied","Data":"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc"} Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.218627 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dbrb2" event={"ID":"eab6fbed-7856-4d82-baa7-5507198362d6","Type":"ContainerDied","Data":"c043ddc4d09754551e0ff50ecd4d4d661d5da5bee3378c0d102aef9ab4cd54e1"} Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.232367 5002 scope.go:117] "RemoveContainer" containerID="8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.246546 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.249469 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dbrb2"] Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.252727 5002 scope.go:117] "RemoveContainer" containerID="b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.261039 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.265148 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-84p8s"] Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.266727 5002 scope.go:117] "RemoveContainer" containerID="8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.267189 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99\": container with ID starting with 8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99 not found: ID does not exist" containerID="8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.267260 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99"} err="failed to get container status \"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99\": rpc error: code = NotFound desc = could not find container \"8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99\": container with ID starting with 8dbb5e903f9942cfb7b4343cd8cc8eba814c9b6f04b0968350c99297cfdcdc99 not found: ID does not exist" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.267341 5002 scope.go:117] "RemoveContainer" containerID="8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.267794 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3\": container with ID starting with 8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3 not found: ID does not exist" containerID="8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.267829 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3"} err="failed to get container status \"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3\": rpc error: code = NotFound desc = could not find container \"8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3\": container with ID starting with 8675ec5a1b8a09a38d6446c73fdab36d3468fd4c5dc6bf45b093e50d7830aff3 not found: ID does not exist" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.267886 5002 scope.go:117] "RemoveContainer" containerID="b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.268493 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4\": container with ID starting with b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4 not found: ID does not exist" containerID="b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.268517 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4"} err="failed to get container status \"b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4\": rpc error: code = NotFound desc = could not find container \"b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4\": container with ID starting with b1951bb26aea09434fc0fd20263d7623181f02de894b90b7d21abda8b6140dd4 not found: ID does not exist" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.268531 5002 scope.go:117] "RemoveContainer" containerID="4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.301735 5002 scope.go:117] "RemoveContainer" containerID="7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.318448 5002 scope.go:117] "RemoveContainer" containerID="b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.346767 5002 scope.go:117] "RemoveContainer" containerID="4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.347260 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc\": container with ID starting with 4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc not found: ID does not exist" containerID="4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.347315 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc"} err="failed to get container status \"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc\": rpc error: code = NotFound desc = could not find container \"4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc\": container with ID starting with 4a484efa9b489e55dfbadab8228852ee64e7041d2196e15df4a245b3853c41fc not found: ID does not exist" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.347353 5002 scope.go:117] "RemoveContainer" containerID="7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.347767 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a\": container with ID starting with 7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a not found: ID does not exist" containerID="7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.347810 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a"} err="failed to get container status \"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a\": rpc error: code = NotFound desc = could not find container \"7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a\": container with ID starting with 7bb3197a36d8ab9994e8fb5773c08471e4cac8e030eb7fa6fc987e17ef60208a not found: ID does not exist" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.347867 5002 scope.go:117] "RemoveContainer" containerID="b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6" Oct 14 07:54:20 crc kubenswrapper[5002]: E1014 07:54:20.348207 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6\": container with ID starting with b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6 not found: ID does not exist" containerID="b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6" Oct 14 07:54:20 crc kubenswrapper[5002]: I1014 07:54:20.348257 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6"} err="failed to get container status \"b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6\": rpc error: code = NotFound desc = could not find container \"b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6\": container with ID starting with b1f75a0e32290a2c8380f00f3bf93ed6e90c7b18a74e6907a849440eb8ca07f6 not found: ID does not exist" Oct 14 07:54:21 crc kubenswrapper[5002]: I1014 07:54:21.725538 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" path="/var/lib/kubelet/pods/3494b1ed-4ee0-4c88-95b7-183d4ed79660/volumes" Oct 14 07:54:21 crc kubenswrapper[5002]: I1014 07:54:21.726438 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" path="/var/lib/kubelet/pods/eab6fbed-7856-4d82-baa7-5507198362d6/volumes" Oct 14 07:54:21 crc kubenswrapper[5002]: I1014 07:54:21.752390 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:54:21 crc kubenswrapper[5002]: I1014 07:54:21.752612 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fq5jf" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="registry-server" containerID="cri-o://6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b" gracePeriod=2 Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.082936 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.099879 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5hhd\" (UniqueName: \"kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd\") pod \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.099959 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities\") pod \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.100022 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content\") pod \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\" (UID: \"cf427bfd-0d50-43b6-9d26-aa11ee3b149a\") " Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.102462 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities" (OuterVolumeSpecName: "utilities") pod "cf427bfd-0d50-43b6-9d26-aa11ee3b149a" (UID: "cf427bfd-0d50-43b6-9d26-aa11ee3b149a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.114874 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf427bfd-0d50-43b6-9d26-aa11ee3b149a" (UID: "cf427bfd-0d50-43b6-9d26-aa11ee3b149a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.149625 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd" (OuterVolumeSpecName: "kube-api-access-r5hhd") pod "cf427bfd-0d50-43b6-9d26-aa11ee3b149a" (UID: "cf427bfd-0d50-43b6-9d26-aa11ee3b149a"). InnerVolumeSpecName "kube-api-access-r5hhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.200991 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5hhd\" (UniqueName: \"kubernetes.io/projected/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-kube-api-access-r5hhd\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.201245 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.201349 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf427bfd-0d50-43b6-9d26-aa11ee3b149a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.231144 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerID="6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b" exitCode=0 Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.231227 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fq5jf" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.231442 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerDied","Data":"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b"} Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.231566 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fq5jf" event={"ID":"cf427bfd-0d50-43b6-9d26-aa11ee3b149a","Type":"ContainerDied","Data":"67a2b3d1354f9aa011f6781848e3eb729c0bbba1f7d19c981ff9d18d40c8ee1e"} Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.231612 5002 scope.go:117] "RemoveContainer" containerID="6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.244883 5002 scope.go:117] "RemoveContainer" containerID="02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.260303 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.262242 5002 scope.go:117] "RemoveContainer" containerID="70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.263328 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fq5jf"] Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.289101 5002 scope.go:117] "RemoveContainer" containerID="6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b" Oct 14 07:54:22 crc kubenswrapper[5002]: E1014 07:54:22.289496 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b\": container with ID starting with 6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b not found: ID does not exist" containerID="6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.289540 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b"} err="failed to get container status \"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b\": rpc error: code = NotFound desc = could not find container \"6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b\": container with ID starting with 6df66cc3a67f5c462fccfe158972e5df0a1161d810238d39804b59fc58eb3b4b not found: ID does not exist" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.289564 5002 scope.go:117] "RemoveContainer" containerID="02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93" Oct 14 07:54:22 crc kubenswrapper[5002]: E1014 07:54:22.289953 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93\": container with ID starting with 02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93 not found: ID does not exist" containerID="02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.289979 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93"} err="failed to get container status \"02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93\": rpc error: code = NotFound desc = could not find container \"02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93\": container with ID starting with 02516caf4f4d8aeb2c2bb06f4fc3873259f16e0e6d04526f471a998d1887ca93 not found: ID does not exist" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.289995 5002 scope.go:117] "RemoveContainer" containerID="70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459" Oct 14 07:54:22 crc kubenswrapper[5002]: E1014 07:54:22.290251 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459\": container with ID starting with 70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459 not found: ID does not exist" containerID="70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459" Oct 14 07:54:22 crc kubenswrapper[5002]: I1014 07:54:22.290275 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459"} err="failed to get container status \"70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459\": rpc error: code = NotFound desc = could not find container \"70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459\": container with ID starting with 70ad177692546decbcd8fc57571e52a6787828dc425950251793321559b04459 not found: ID does not exist" Oct 14 07:54:23 crc kubenswrapper[5002]: I1014 07:54:23.726391 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" path="/var/lib/kubelet/pods/cf427bfd-0d50-43b6-9d26-aa11ee3b149a/volumes" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.153329 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.153878 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dq285" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="registry-server" containerID="cri-o://4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22" gracePeriod=2 Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.552345 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.731782 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities\") pod \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.731904 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5ns9\" (UniqueName: \"kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9\") pod \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.731960 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content\") pod \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\" (UID: \"b40f0b2e-9c3a-4f91-acd6-486a233cee10\") " Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.734381 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities" (OuterVolumeSpecName: "utilities") pod "b40f0b2e-9c3a-4f91-acd6-486a233cee10" (UID: "b40f0b2e-9c3a-4f91-acd6-486a233cee10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.743306 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9" (OuterVolumeSpecName: "kube-api-access-b5ns9") pod "b40f0b2e-9c3a-4f91-acd6-486a233cee10" (UID: "b40f0b2e-9c3a-4f91-acd6-486a233cee10"). InnerVolumeSpecName "kube-api-access-b5ns9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.821033 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b40f0b2e-9c3a-4f91-acd6-486a233cee10" (UID: "b40f0b2e-9c3a-4f91-acd6-486a233cee10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.832982 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5ns9\" (UniqueName: \"kubernetes.io/projected/b40f0b2e-9c3a-4f91-acd6-486a233cee10-kube-api-access-b5ns9\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.833136 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:24 crc kubenswrapper[5002]: I1014 07:54:24.833223 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40f0b2e-9c3a-4f91-acd6-486a233cee10-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.247247 5002 generic.go:334] "Generic (PLEG): container finished" podID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerID="4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22" exitCode=0 Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.247293 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerDied","Data":"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22"} Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.247320 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dq285" event={"ID":"b40f0b2e-9c3a-4f91-acd6-486a233cee10","Type":"ContainerDied","Data":"cb58630f27289c8c8bc46c28faf72cfbf33e5ab8710544142474e5a000258ade"} Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.247340 5002 scope.go:117] "RemoveContainer" containerID="4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.247461 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dq285" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.267598 5002 scope.go:117] "RemoveContainer" containerID="5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.277488 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.280550 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dq285"] Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.303189 5002 scope.go:117] "RemoveContainer" containerID="8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.326243 5002 scope.go:117] "RemoveContainer" containerID="4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22" Oct 14 07:54:25 crc kubenswrapper[5002]: E1014 07:54:25.326826 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22\": container with ID starting with 4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22 not found: ID does not exist" containerID="4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.326940 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22"} err="failed to get container status \"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22\": rpc error: code = NotFound desc = could not find container \"4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22\": container with ID starting with 4e5fdad0d5f4f6c22bd657e2029338bbcde4c831598715ae42b471f0f2273b22 not found: ID does not exist" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.327038 5002 scope.go:117] "RemoveContainer" containerID="5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321" Oct 14 07:54:25 crc kubenswrapper[5002]: E1014 07:54:25.327477 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321\": container with ID starting with 5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321 not found: ID does not exist" containerID="5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.327513 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321"} err="failed to get container status \"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321\": rpc error: code = NotFound desc = could not find container \"5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321\": container with ID starting with 5ad46cea9594c62cd078b0d7b73ce3d2419fbd2311184a641b76de68a327f321 not found: ID does not exist" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.327542 5002 scope.go:117] "RemoveContainer" containerID="8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a" Oct 14 07:54:25 crc kubenswrapper[5002]: E1014 07:54:25.327888 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a\": container with ID starting with 8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a not found: ID does not exist" containerID="8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.327972 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a"} err="failed to get container status \"8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a\": rpc error: code = NotFound desc = could not find container \"8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a\": container with ID starting with 8fd82aa23a1d2d2422e92c9d9c9ac64723e4d74dc338f7a3e05658789a5b136a not found: ID does not exist" Oct 14 07:54:25 crc kubenswrapper[5002]: I1014 07:54:25.726441 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" path="/var/lib/kubelet/pods/b40f0b2e-9c3a-4f91-acd6-486a233cee10/volumes" Oct 14 07:54:39 crc kubenswrapper[5002]: I1014 07:54:39.218133 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:54:39 crc kubenswrapper[5002]: I1014 07:54:39.218747 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:54:39 crc kubenswrapper[5002]: I1014 07:54:39.218825 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:54:39 crc kubenswrapper[5002]: I1014 07:54:39.219637 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:54:39 crc kubenswrapper[5002]: I1014 07:54:39.219783 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b" gracePeriod=600 Oct 14 07:54:40 crc kubenswrapper[5002]: I1014 07:54:40.330558 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b" exitCode=0 Oct 14 07:54:40 crc kubenswrapper[5002]: I1014 07:54:40.330686 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b"} Oct 14 07:54:40 crc kubenswrapper[5002]: I1014 07:54:40.331136 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf"} Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.922138 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.922819 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9xr4f" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="registry-server" containerID="cri-o://c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5" gracePeriod=30 Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.932045 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.932342 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cqzkm" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="registry-server" containerID="cri-o://054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f" gracePeriod=30 Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.948262 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.948565 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" containerID="cri-o://e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97" gracePeriod=30 Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.960954 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.961229 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jlfz8" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="registry-server" containerID="cri-o://d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98" gracePeriod=30 Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.968385 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.969170 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pgl6v" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="registry-server" containerID="cri-o://d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269" gracePeriod=30 Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.971950 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cq8mp"] Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972179 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972197 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972208 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972217 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972234 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b09b338a-19a0-4b68-9774-aeb15f68e63c" containerName="pruner" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972243 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b09b338a-19a0-4b68-9774-aeb15f68e63c" containerName="pruner" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972254 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972262 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972271 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972280 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972292 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972301 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972311 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972319 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972332 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f63d16a-af72-4d47-828d-8ac2532a6196" containerName="collect-profiles" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972341 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f63d16a-af72-4d47-828d-8ac2532a6196" containerName="collect-profiles" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972352 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972361 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972374 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972382 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972395 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972404 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972416 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972425 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="extract-content" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972436 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972446 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="extract-utilities" Oct 14 07:55:09 crc kubenswrapper[5002]: E1014 07:55:09.972717 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972730 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972872 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b09b338a-19a0-4b68-9774-aeb15f68e63c" containerName="pruner" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972886 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3494b1ed-4ee0-4c88-95b7-183d4ed79660" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972903 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40f0b2e-9c3a-4f91-acd6-486a233cee10" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972917 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf427bfd-0d50-43b6-9d26-aa11ee3b149a" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972928 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab6fbed-7856-4d82-baa7-5507198362d6" containerName="registry-server" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.972941 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f63d16a-af72-4d47-828d-8ac2532a6196" containerName="collect-profiles" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.973594 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:09 crc kubenswrapper[5002]: I1014 07:55:09.982337 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cq8mp"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.005176 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.005222 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rbzk\" (UniqueName: \"kubernetes.io/projected/651e3f67-0cad-4654-8c98-0672dd56f367-kube-api-access-2rbzk\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.005324 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.106695 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.106744 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.106766 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rbzk\" (UniqueName: \"kubernetes.io/projected/651e3f67-0cad-4654-8c98-0672dd56f367-kube-api-access-2rbzk\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.110177 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.117753 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/651e3f67-0cad-4654-8c98-0672dd56f367-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.125953 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rbzk\" (UniqueName: \"kubernetes.io/projected/651e3f67-0cad-4654-8c98-0672dd56f367-kube-api-access-2rbzk\") pod \"marketplace-operator-79b997595-cq8mp\" (UID: \"651e3f67-0cad-4654-8c98-0672dd56f367\") " pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.337977 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.352754 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.368480 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.397269 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410550 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities\") pod \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410598 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk28k\" (UniqueName: \"kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k\") pod \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410640 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zc677\" (UniqueName: \"kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677\") pod \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410668 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content\") pod \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410702 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities\") pod \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\" (UID: \"36a9b93e-3378-4afd-8f4d-cd4f99d4f430\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.410726 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content\") pod \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\" (UID: \"736b4c73-c83f-4bf9-9e6f-1b468f3f9996\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.411165 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.411868 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities" (OuterVolumeSpecName: "utilities") pod "36a9b93e-3378-4afd-8f4d-cd4f99d4f430" (UID: "36a9b93e-3378-4afd-8f4d-cd4f99d4f430"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.412166 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities" (OuterVolumeSpecName: "utilities") pod "736b4c73-c83f-4bf9-9e6f-1b468f3f9996" (UID: "736b4c73-c83f-4bf9-9e6f-1b468f3f9996"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.418701 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677" (OuterVolumeSpecName: "kube-api-access-zc677") pod "736b4c73-c83f-4bf9-9e6f-1b468f3f9996" (UID: "736b4c73-c83f-4bf9-9e6f-1b468f3f9996"). InnerVolumeSpecName "kube-api-access-zc677". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.423959 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k" (OuterVolumeSpecName: "kube-api-access-kk28k") pod "36a9b93e-3378-4afd-8f4d-cd4f99d4f430" (UID: "36a9b93e-3378-4afd-8f4d-cd4f99d4f430"). InnerVolumeSpecName "kube-api-access-kk28k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.428531 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.492220 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36a9b93e-3378-4afd-8f4d-cd4f99d4f430" (UID: "36a9b93e-3378-4afd-8f4d-cd4f99d4f430"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513124 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg96l\" (UniqueName: \"kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l\") pod \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513200 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities\") pod \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513233 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content\") pod \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513292 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics\") pod \"dd63d35e-85fa-44ae-9389-28b816cffae3\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content\") pod \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513343 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities\") pod \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\" (UID: \"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513377 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2rjk\" (UniqueName: \"kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk\") pod \"dd63d35e-85fa-44ae-9389-28b816cffae3\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513394 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca\") pod \"dd63d35e-85fa-44ae-9389-28b816cffae3\" (UID: \"dd63d35e-85fa-44ae-9389-28b816cffae3\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513417 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxvp4\" (UniqueName: \"kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4\") pod \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\" (UID: \"d40e9f14-1956-43e3-a22c-4fcc1e5f6450\") " Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513640 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513658 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513668 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513677 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk28k\" (UniqueName: \"kubernetes.io/projected/36a9b93e-3378-4afd-8f4d-cd4f99d4f430-kube-api-access-kk28k\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.513687 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zc677\" (UniqueName: \"kubernetes.io/projected/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-kube-api-access-zc677\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.517963 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities" (OuterVolumeSpecName: "utilities") pod "d40e9f14-1956-43e3-a22c-4fcc1e5f6450" (UID: "d40e9f14-1956-43e3-a22c-4fcc1e5f6450"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.521313 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities" (OuterVolumeSpecName: "utilities") pod "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" (UID: "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.530134 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk" (OuterVolumeSpecName: "kube-api-access-f2rjk") pod "dd63d35e-85fa-44ae-9389-28b816cffae3" (UID: "dd63d35e-85fa-44ae-9389-28b816cffae3"). InnerVolumeSpecName "kube-api-access-f2rjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.530412 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dd63d35e-85fa-44ae-9389-28b816cffae3" (UID: "dd63d35e-85fa-44ae-9389-28b816cffae3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.530819 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dd63d35e-85fa-44ae-9389-28b816cffae3" (UID: "dd63d35e-85fa-44ae-9389-28b816cffae3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.532258 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4" (OuterVolumeSpecName: "kube-api-access-sxvp4") pod "d40e9f14-1956-43e3-a22c-4fcc1e5f6450" (UID: "d40e9f14-1956-43e3-a22c-4fcc1e5f6450"). InnerVolumeSpecName "kube-api-access-sxvp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.532662 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "736b4c73-c83f-4bf9-9e6f-1b468f3f9996" (UID: "736b4c73-c83f-4bf9-9e6f-1b468f3f9996"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.543738 5002 generic.go:334] "Generic (PLEG): container finished" podID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerID="c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5" exitCode=0 Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.543814 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerDied","Data":"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.543859 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9xr4f" event={"ID":"736b4c73-c83f-4bf9-9e6f-1b468f3f9996","Type":"ContainerDied","Data":"e3549941f7be81c945457badd004a5de1d2036800041bec6386e9f2aa03a92ce"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.543895 5002 scope.go:117] "RemoveContainer" containerID="c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.544010 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9xr4f" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.554305 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" (UID: "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.554488 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l" (OuterVolumeSpecName: "kube-api-access-dg96l") pod "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" (UID: "45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6"). InnerVolumeSpecName "kube-api-access-dg96l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.556429 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.564806 5002 generic.go:334] "Generic (PLEG): container finished" podID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerID="d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98" exitCode=0 Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.565109 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jlfz8" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.566605 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerDied","Data":"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.566640 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jlfz8" event={"ID":"45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6","Type":"ContainerDied","Data":"02fa4e0cba546e1b1e1f7b344538fcfc29712a714b58aae8c6113b77978737a3"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.576965 5002 generic.go:334] "Generic (PLEG): container finished" podID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerID="d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269" exitCode=0 Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.577027 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerDied","Data":"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.577059 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgl6v" event={"ID":"d40e9f14-1956-43e3-a22c-4fcc1e5f6450","Type":"ContainerDied","Data":"c7031efc41bc111e428e70737cc10a3784eeb9b614253199d695c06956d7c15b"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.577127 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgl6v" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.578899 5002 generic.go:334] "Generic (PLEG): container finished" podID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerID="054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f" exitCode=0 Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.578936 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerDied","Data":"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.578951 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqzkm" event={"ID":"36a9b93e-3378-4afd-8f4d-cd4f99d4f430","Type":"ContainerDied","Data":"d3ef6124fba4ec3e1ae5fd78df81d2fbaa1c83177a948836afb0123817251d84"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.579005 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqzkm" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.590445 5002 generic.go:334] "Generic (PLEG): container finished" podID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerID="e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97" exitCode=0 Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.590479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" event={"ID":"dd63d35e-85fa-44ae-9389-28b816cffae3","Type":"ContainerDied","Data":"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.590505 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" event={"ID":"dd63d35e-85fa-44ae-9389-28b816cffae3","Type":"ContainerDied","Data":"bb693d7c92d1c067952113bb2be8b9fb76ca5b195ecda44a58877d0a9d7e6bf6"} Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.590572 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ndpjk" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.597436 5002 scope.go:117] "RemoveContainer" containerID="c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.600285 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.602411 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9xr4f"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614470 5002 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614503 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614513 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2rjk\" (UniqueName: \"kubernetes.io/projected/dd63d35e-85fa-44ae-9389-28b816cffae3-kube-api-access-f2rjk\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614521 5002 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dd63d35e-85fa-44ae-9389-28b816cffae3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614530 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxvp4\" (UniqueName: \"kubernetes.io/projected/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-kube-api-access-sxvp4\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614565 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg96l\" (UniqueName: \"kubernetes.io/projected/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-kube-api-access-dg96l\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614573 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/736b4c73-c83f-4bf9-9e6f-1b468f3f9996-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614582 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.614589 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.630698 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.634326 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jlfz8"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.639084 5002 scope.go:117] "RemoveContainer" containerID="33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.651692 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d40e9f14-1956-43e3-a22c-4fcc1e5f6450" (UID: "d40e9f14-1956-43e3-a22c-4fcc1e5f6450"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.657725 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.660025 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cqzkm"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.662151 5002 scope.go:117] "RemoveContainer" containerID="c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.662418 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5\": container with ID starting with c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5 not found: ID does not exist" containerID="c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.662449 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5"} err="failed to get container status \"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5\": rpc error: code = NotFound desc = could not find container \"c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5\": container with ID starting with c1f4d70d14b19c64cba1f19a0d02852af1c553a0f957402945de836e87658ee5 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.662468 5002 scope.go:117] "RemoveContainer" containerID="c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.662665 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171\": container with ID starting with c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171 not found: ID does not exist" containerID="c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.662726 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171"} err="failed to get container status \"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171\": rpc error: code = NotFound desc = could not find container \"c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171\": container with ID starting with c5ebae3236352fac522044ffcf87269fbcba10199d52587ef0ffefa00f68e171 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.662757 5002 scope.go:117] "RemoveContainer" containerID="33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.663395 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8\": container with ID starting with 33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8 not found: ID does not exist" containerID="33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.663449 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8"} err="failed to get container status \"33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8\": rpc error: code = NotFound desc = could not find container \"33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8\": container with ID starting with 33d3881270f92752410182e89e1f5a7bfba861d35a94d184ab3361b46cb921e8 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.663531 5002 scope.go:117] "RemoveContainer" containerID="d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.679025 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.691993 5002 scope.go:117] "RemoveContainer" containerID="a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.692602 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ndpjk"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.716734 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d40e9f14-1956-43e3-a22c-4fcc1e5f6450-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.745995 5002 scope.go:117] "RemoveContainer" containerID="7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.768745 5002 scope.go:117] "RemoveContainer" containerID="d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.772063 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98\": container with ID starting with d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98 not found: ID does not exist" containerID="d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.772110 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98"} err="failed to get container status \"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98\": rpc error: code = NotFound desc = could not find container \"d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98\": container with ID starting with d4cdb007bfc4f7575826504ce646e270d5966e3ed360dc5357e8ff2a50e72f98 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.772137 5002 scope.go:117] "RemoveContainer" containerID="a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.775327 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a\": container with ID starting with a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a not found: ID does not exist" containerID="a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.775361 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a"} err="failed to get container status \"a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a\": rpc error: code = NotFound desc = could not find container \"a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a\": container with ID starting with a387127feb65ad41ef2899edb30ad727e3c68deb3e5422315729a5a611206a7a not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.775386 5002 scope.go:117] "RemoveContainer" containerID="7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.775985 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610\": container with ID starting with 7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610 not found: ID does not exist" containerID="7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.776013 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610"} err="failed to get container status \"7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610\": rpc error: code = NotFound desc = could not find container \"7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610\": container with ID starting with 7b45a0e512659d7317dee14b089eea6194f00d29f1e7ab7ce5cb166f38d60610 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.776028 5002 scope.go:117] "RemoveContainer" containerID="d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.786786 5002 scope.go:117] "RemoveContainer" containerID="bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.810512 5002 scope.go:117] "RemoveContainer" containerID="2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.824404 5002 scope.go:117] "RemoveContainer" containerID="d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.824911 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269\": container with ID starting with d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269 not found: ID does not exist" containerID="d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.824955 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269"} err="failed to get container status \"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269\": rpc error: code = NotFound desc = could not find container \"d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269\": container with ID starting with d6015daa6df11cf8509284527a7555f5aeb4e71b1d06d3c60f11e45428d7a269 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.824989 5002 scope.go:117] "RemoveContainer" containerID="bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.825386 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67\": container with ID starting with bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67 not found: ID does not exist" containerID="bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.825429 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67"} err="failed to get container status \"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67\": rpc error: code = NotFound desc = could not find container \"bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67\": container with ID starting with bc0fe999b996c6651054c6da494a549d6507da908354c4e5f6f319264430de67 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.825459 5002 scope.go:117] "RemoveContainer" containerID="2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.827060 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540\": container with ID starting with 2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540 not found: ID does not exist" containerID="2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.827086 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540"} err="failed to get container status \"2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540\": rpc error: code = NotFound desc = could not find container \"2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540\": container with ID starting with 2e03783ba857bdc0c906618344359dccddd27037cc9acf73ba549d4a0fba7540 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.827106 5002 scope.go:117] "RemoveContainer" containerID="054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.838810 5002 scope.go:117] "RemoveContainer" containerID="67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.850384 5002 scope.go:117] "RemoveContainer" containerID="26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.867353 5002 scope.go:117] "RemoveContainer" containerID="054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.871297 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f\": container with ID starting with 054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f not found: ID does not exist" containerID="054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.871331 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f"} err="failed to get container status \"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f\": rpc error: code = NotFound desc = could not find container \"054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f\": container with ID starting with 054ac7dfc2ec761d8062594b43ab2ea83a928fec1e2fc5492ff9ddab06f2ea9f not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.871360 5002 scope.go:117] "RemoveContainer" containerID="67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.871770 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396\": container with ID starting with 67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396 not found: ID does not exist" containerID="67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.871793 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396"} err="failed to get container status \"67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396\": rpc error: code = NotFound desc = could not find container \"67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396\": container with ID starting with 67f7a224aca0255897708d0861fb9733da551fc284b0abfd2962f0df9b0de396 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.871807 5002 scope.go:117] "RemoveContainer" containerID="26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.872162 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541\": container with ID starting with 26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541 not found: ID does not exist" containerID="26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.872184 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541"} err="failed to get container status \"26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541\": rpc error: code = NotFound desc = could not find container \"26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541\": container with ID starting with 26c3e2d11c6de1440f4eaf43756b79e01d9373c0213dc7dea46a747f49d8a541 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.872197 5002 scope.go:117] "RemoveContainer" containerID="e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.891624 5002 scope.go:117] "RemoveContainer" containerID="e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97" Oct 14 07:55:10 crc kubenswrapper[5002]: E1014 07:55:10.892307 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97\": container with ID starting with e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97 not found: ID does not exist" containerID="e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.892334 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97"} err="failed to get container status \"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97\": rpc error: code = NotFound desc = could not find container \"e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97\": container with ID starting with e1325e864fc9a747115aba3c86c68a20bf01844f7e41930f7d2c7f0f66a25d97 not found: ID does not exist" Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.899792 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.907976 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cq8mp"] Oct 14 07:55:10 crc kubenswrapper[5002]: I1014 07:55:10.910152 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pgl6v"] Oct 14 07:55:10 crc kubenswrapper[5002]: W1014 07:55:10.924627 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod651e3f67_0cad_4654_8c98_0672dd56f367.slice/crio-e7aa14113d2648494d7e7e05da6b11b22738470ec21660d078d2582c87b3888f WatchSource:0}: Error finding container e7aa14113d2648494d7e7e05da6b11b22738470ec21660d078d2582c87b3888f: Status 404 returned error can't find the container with id e7aa14113d2648494d7e7e05da6b11b22738470ec21660d078d2582c87b3888f Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.601273 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" event={"ID":"651e3f67-0cad-4654-8c98-0672dd56f367","Type":"ContainerStarted","Data":"b6496184e2b252f6fa39b85835a463de13e63d34302a44a6f24bccf4a9877441"} Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.601607 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.601618 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" event={"ID":"651e3f67-0cad-4654-8c98-0672dd56f367","Type":"ContainerStarted","Data":"e7aa14113d2648494d7e7e05da6b11b22738470ec21660d078d2582c87b3888f"} Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.606045 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.621988 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cq8mp" podStartSLOduration=2.6219637159999998 podStartE2EDuration="2.621963716s" podCreationTimestamp="2025-10-14 07:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:55:11.620440414 +0000 UTC m=+244.601679916" watchObservedRunningTime="2025-10-14 07:55:11.621963716 +0000 UTC m=+244.603203208" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.725907 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" path="/var/lib/kubelet/pods/36a9b93e-3378-4afd-8f4d-cd4f99d4f430/volumes" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.726470 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" path="/var/lib/kubelet/pods/45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6/volumes" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.727038 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" path="/var/lib/kubelet/pods/736b4c73-c83f-4bf9-9e6f-1b468f3f9996/volumes" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.727963 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" path="/var/lib/kubelet/pods/d40e9f14-1956-43e3-a22c-4fcc1e5f6450/volumes" Oct 14 07:55:11 crc kubenswrapper[5002]: I1014 07:55:11.728654 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" path="/var/lib/kubelet/pods/dd63d35e-85fa-44ae-9389-28b816cffae3/volumes" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.149568 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nggtm"] Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.149829 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.149885 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.149905 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.149918 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.149942 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.149956 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.149970 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.149982 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.149997 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150008 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150025 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150037 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150050 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150062 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150082 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150094 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150110 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150122 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150137 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150149 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150164 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150176 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="extract-utilities" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150195 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150207 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: E1014 07:55:12.150225 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150236 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="extract-content" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150389 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="45a1e8b6-604b-41b2-b54a-1e2fdf8ef3d6" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150418 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="36a9b93e-3378-4afd-8f4d-cd4f99d4f430" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150434 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd63d35e-85fa-44ae-9389-28b816cffae3" containerName="marketplace-operator" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150451 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="736b4c73-c83f-4bf9-9e6f-1b468f3f9996" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.150468 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40e9f14-1956-43e3-a22c-4fcc1e5f6450" containerName="registry-server" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.152461 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.156025 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.156990 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nggtm"] Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.232048 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-utilities\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.232089 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt4mp\" (UniqueName: \"kubernetes.io/projected/7f04708c-e386-4cfa-979f-9b930ee896ae-kube-api-access-nt4mp\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.232126 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-catalog-content\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.335545 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-utilities\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.335600 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt4mp\" (UniqueName: \"kubernetes.io/projected/7f04708c-e386-4cfa-979f-9b930ee896ae-kube-api-access-nt4mp\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.335632 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-catalog-content\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.336042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-catalog-content\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.336140 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f04708c-e386-4cfa-979f-9b930ee896ae-utilities\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.339381 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ts5dh"] Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.340224 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.342164 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.352697 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ts5dh"] Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.370994 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt4mp\" (UniqueName: \"kubernetes.io/projected/7f04708c-e386-4cfa-979f-9b930ee896ae-kube-api-access-nt4mp\") pod \"redhat-marketplace-nggtm\" (UID: \"7f04708c-e386-4cfa-979f-9b930ee896ae\") " pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.436697 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-catalog-content\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.436741 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5gmq\" (UniqueName: \"kubernetes.io/projected/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-kube-api-access-j5gmq\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.436776 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-utilities\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.477477 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.537604 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-catalog-content\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.538003 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5gmq\" (UniqueName: \"kubernetes.io/projected/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-kube-api-access-j5gmq\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.538056 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-utilities\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.538187 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-catalog-content\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.538474 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-utilities\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.574461 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5gmq\" (UniqueName: \"kubernetes.io/projected/ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15-kube-api-access-j5gmq\") pod \"redhat-operators-ts5dh\" (UID: \"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15\") " pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.665680 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nggtm"] Oct 14 07:55:12 crc kubenswrapper[5002]: I1014 07:55:12.672972 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:12 crc kubenswrapper[5002]: W1014 07:55:12.675240 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f04708c_e386_4cfa_979f_9b930ee896ae.slice/crio-318930d687203455ad552fa6426c70075c21f298e619b447f5f8961ba1320fd3 WatchSource:0}: Error finding container 318930d687203455ad552fa6426c70075c21f298e619b447f5f8961ba1320fd3: Status 404 returned error can't find the container with id 318930d687203455ad552fa6426c70075c21f298e619b447f5f8961ba1320fd3 Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.077051 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ts5dh"] Oct 14 07:55:13 crc kubenswrapper[5002]: W1014 07:55:13.082166 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae36f67a_ccd9_47f8_8030_6e1eb0fbcb15.slice/crio-828b9e4cddd8b543b9e2a39cd21cdc8ea1964427249d2f0a3b54b04461e492c7 WatchSource:0}: Error finding container 828b9e4cddd8b543b9e2a39cd21cdc8ea1964427249d2f0a3b54b04461e492c7: Status 404 returned error can't find the container with id 828b9e4cddd8b543b9e2a39cd21cdc8ea1964427249d2f0a3b54b04461e492c7 Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.618311 5002 generic.go:334] "Generic (PLEG): container finished" podID="ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15" containerID="6eb7d12fe1de7a12dc60de2d73ddf9ef1a1c12a9023464fe34d48c9edc66dd7d" exitCode=0 Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.618377 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ts5dh" event={"ID":"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15","Type":"ContainerDied","Data":"6eb7d12fe1de7a12dc60de2d73ddf9ef1a1c12a9023464fe34d48c9edc66dd7d"} Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.618608 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ts5dh" event={"ID":"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15","Type":"ContainerStarted","Data":"828b9e4cddd8b543b9e2a39cd21cdc8ea1964427249d2f0a3b54b04461e492c7"} Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.623631 5002 generic.go:334] "Generic (PLEG): container finished" podID="7f04708c-e386-4cfa-979f-9b930ee896ae" containerID="d230ca68067b6a2ece19b96d9b656f2dc3b8ceded51c443ee64d36e635a9fc4e" exitCode=0 Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.624163 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nggtm" event={"ID":"7f04708c-e386-4cfa-979f-9b930ee896ae","Type":"ContainerDied","Data":"d230ca68067b6a2ece19b96d9b656f2dc3b8ceded51c443ee64d36e635a9fc4e"} Oct 14 07:55:13 crc kubenswrapper[5002]: I1014 07:55:13.624197 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nggtm" event={"ID":"7f04708c-e386-4cfa-979f-9b930ee896ae","Type":"ContainerStarted","Data":"318930d687203455ad552fa6426c70075c21f298e619b447f5f8961ba1320fd3"} Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.552758 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6wmt"] Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.554752 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.556395 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.564215 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6wmt"] Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.629146 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ts5dh" event={"ID":"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15","Type":"ContainerStarted","Data":"4a3d529df1a398a076366ebea22661e232119b161aeb33881322314c2cd18e80"} Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.633946 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nggtm" event={"ID":"7f04708c-e386-4cfa-979f-9b930ee896ae","Type":"ContainerStarted","Data":"40db93eeac54f35e4b526f0eb500847761ec9a978ffe37f157dc07ab55ac5ec9"} Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.670388 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-catalog-content\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.670723 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdgh6\" (UniqueName: \"kubernetes.io/projected/1d5a10a8-19fd-46f5-8062-d3e1db89f218-kube-api-access-gdgh6\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.670771 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-utilities\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.743636 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dzd5b"] Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.747578 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.750475 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.755263 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dzd5b"] Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.771754 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-catalog-content\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.771884 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-catalog-content\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.771921 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdgh6\" (UniqueName: \"kubernetes.io/projected/1d5a10a8-19fd-46f5-8062-d3e1db89f218-kube-api-access-gdgh6\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.771949 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-utilities\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.771978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxb4f\" (UniqueName: \"kubernetes.io/projected/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-kube-api-access-pxb4f\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.772016 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-utilities\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.772530 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-catalog-content\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.773127 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d5a10a8-19fd-46f5-8062-d3e1db89f218-utilities\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.793574 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdgh6\" (UniqueName: \"kubernetes.io/projected/1d5a10a8-19fd-46f5-8062-d3e1db89f218-kube-api-access-gdgh6\") pod \"community-operators-v6wmt\" (UID: \"1d5a10a8-19fd-46f5-8062-d3e1db89f218\") " pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.873621 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-catalog-content\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.873703 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxb4f\" (UniqueName: \"kubernetes.io/projected/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-kube-api-access-pxb4f\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.873737 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-utilities\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.874120 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-utilities\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.874325 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-catalog-content\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.882433 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:14 crc kubenswrapper[5002]: I1014 07:55:14.894796 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxb4f\" (UniqueName: \"kubernetes.io/projected/1cb8dd4f-12fd-46d1-99d3-61b66d6147a1-kube-api-access-pxb4f\") pod \"certified-operators-dzd5b\" (UID: \"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1\") " pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.067082 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.307337 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6wmt"] Oct 14 07:55:15 crc kubenswrapper[5002]: W1014 07:55:15.312108 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d5a10a8_19fd_46f5_8062_d3e1db89f218.slice/crio-d7fbf9995270a7d0c16f5412fb2548f73f18c38e03c9224d09d1679793922c0f WatchSource:0}: Error finding container d7fbf9995270a7d0c16f5412fb2548f73f18c38e03c9224d09d1679793922c0f: Status 404 returned error can't find the container with id d7fbf9995270a7d0c16f5412fb2548f73f18c38e03c9224d09d1679793922c0f Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.452219 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dzd5b"] Oct 14 07:55:15 crc kubenswrapper[5002]: W1014 07:55:15.465558 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cb8dd4f_12fd_46d1_99d3_61b66d6147a1.slice/crio-c9ed37492eeed3f9351ce9a35bcbbb6d775312bdcc6928e5e63daf714fcaa5b5 WatchSource:0}: Error finding container c9ed37492eeed3f9351ce9a35bcbbb6d775312bdcc6928e5e63daf714fcaa5b5: Status 404 returned error can't find the container with id c9ed37492eeed3f9351ce9a35bcbbb6d775312bdcc6928e5e63daf714fcaa5b5 Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.640658 5002 generic.go:334] "Generic (PLEG): container finished" podID="ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15" containerID="4a3d529df1a398a076366ebea22661e232119b161aeb33881322314c2cd18e80" exitCode=0 Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.640734 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ts5dh" event={"ID":"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15","Type":"ContainerDied","Data":"4a3d529df1a398a076366ebea22661e232119b161aeb33881322314c2cd18e80"} Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.650268 5002 generic.go:334] "Generic (PLEG): container finished" podID="7f04708c-e386-4cfa-979f-9b930ee896ae" containerID="40db93eeac54f35e4b526f0eb500847761ec9a978ffe37f157dc07ab55ac5ec9" exitCode=0 Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.650401 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nggtm" event={"ID":"7f04708c-e386-4cfa-979f-9b930ee896ae","Type":"ContainerDied","Data":"40db93eeac54f35e4b526f0eb500847761ec9a978ffe37f157dc07ab55ac5ec9"} Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.660402 5002 generic.go:334] "Generic (PLEG): container finished" podID="1d5a10a8-19fd-46f5-8062-d3e1db89f218" containerID="c08a87e84304896682cc9124c0388a2f8ccba39d0ff00658b28c685b42341e25" exitCode=0 Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.660498 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6wmt" event={"ID":"1d5a10a8-19fd-46f5-8062-d3e1db89f218","Type":"ContainerDied","Data":"c08a87e84304896682cc9124c0388a2f8ccba39d0ff00658b28c685b42341e25"} Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.660526 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6wmt" event={"ID":"1d5a10a8-19fd-46f5-8062-d3e1db89f218","Type":"ContainerStarted","Data":"d7fbf9995270a7d0c16f5412fb2548f73f18c38e03c9224d09d1679793922c0f"} Oct 14 07:55:15 crc kubenswrapper[5002]: I1014 07:55:15.662064 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzd5b" event={"ID":"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1","Type":"ContainerStarted","Data":"c9ed37492eeed3f9351ce9a35bcbbb6d775312bdcc6928e5e63daf714fcaa5b5"} Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.670473 5002 generic.go:334] "Generic (PLEG): container finished" podID="1d5a10a8-19fd-46f5-8062-d3e1db89f218" containerID="91e71d93ab468d5e2028ee695c80e1298ea7d1f094067c113f0a7b091b6de17c" exitCode=0 Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.670817 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6wmt" event={"ID":"1d5a10a8-19fd-46f5-8062-d3e1db89f218","Type":"ContainerDied","Data":"91e71d93ab468d5e2028ee695c80e1298ea7d1f094067c113f0a7b091b6de17c"} Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.673819 5002 generic.go:334] "Generic (PLEG): container finished" podID="1cb8dd4f-12fd-46d1-99d3-61b66d6147a1" containerID="abc0cb7c41b34f4c467f8b11f29767a878df1ec47b6d5d2fc26815fef838fcc8" exitCode=0 Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.673913 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzd5b" event={"ID":"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1","Type":"ContainerDied","Data":"abc0cb7c41b34f4c467f8b11f29767a878df1ec47b6d5d2fc26815fef838fcc8"} Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.681501 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ts5dh" event={"ID":"ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15","Type":"ContainerStarted","Data":"dc13d06e401d6568a25b70b401f820d6c7eae2904988cdb1af894eb5c5f9a2e2"} Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.686366 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nggtm" event={"ID":"7f04708c-e386-4cfa-979f-9b930ee896ae","Type":"ContainerStarted","Data":"a7ae639d8c2ba22acd00adc5d379d4396d9dbbb9caa455fb209bec64848850ce"} Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.732157 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nggtm" podStartSLOduration=2.154857231 podStartE2EDuration="4.732141413s" podCreationTimestamp="2025-10-14 07:55:12 +0000 UTC" firstStartedPulling="2025-10-14 07:55:13.629221485 +0000 UTC m=+246.610460937" lastFinishedPulling="2025-10-14 07:55:16.206505657 +0000 UTC m=+249.187745119" observedRunningTime="2025-10-14 07:55:16.73095753 +0000 UTC m=+249.712196992" watchObservedRunningTime="2025-10-14 07:55:16.732141413 +0000 UTC m=+249.713380865" Oct 14 07:55:16 crc kubenswrapper[5002]: I1014 07:55:16.755390 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ts5dh" podStartSLOduration=2.2567312409999998 podStartE2EDuration="4.755375684s" podCreationTimestamp="2025-10-14 07:55:12 +0000 UTC" firstStartedPulling="2025-10-14 07:55:13.619871691 +0000 UTC m=+246.601111263" lastFinishedPulling="2025-10-14 07:55:16.118516244 +0000 UTC m=+249.099755706" observedRunningTime="2025-10-14 07:55:16.754235393 +0000 UTC m=+249.735474845" watchObservedRunningTime="2025-10-14 07:55:16.755375684 +0000 UTC m=+249.736615136" Oct 14 07:55:18 crc kubenswrapper[5002]: I1014 07:55:18.702094 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6wmt" event={"ID":"1d5a10a8-19fd-46f5-8062-d3e1db89f218","Type":"ContainerStarted","Data":"727506f58d87183b5e2787317f857b709e10d91dc646db6dbcc380b15978a119"} Oct 14 07:55:18 crc kubenswrapper[5002]: I1014 07:55:18.704531 5002 generic.go:334] "Generic (PLEG): container finished" podID="1cb8dd4f-12fd-46d1-99d3-61b66d6147a1" containerID="b99c1de025a89a03c57ced063ddcff68cd7775ce67b1dc9e005c415bbe596838" exitCode=0 Oct 14 07:55:18 crc kubenswrapper[5002]: I1014 07:55:18.704559 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzd5b" event={"ID":"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1","Type":"ContainerDied","Data":"b99c1de025a89a03c57ced063ddcff68cd7775ce67b1dc9e005c415bbe596838"} Oct 14 07:55:18 crc kubenswrapper[5002]: I1014 07:55:18.724465 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6wmt" podStartSLOduration=3.284817904 podStartE2EDuration="4.724449466s" podCreationTimestamp="2025-10-14 07:55:14 +0000 UTC" firstStartedPulling="2025-10-14 07:55:15.662621296 +0000 UTC m=+248.643860748" lastFinishedPulling="2025-10-14 07:55:17.102252858 +0000 UTC m=+250.083492310" observedRunningTime="2025-10-14 07:55:18.720723264 +0000 UTC m=+251.701962736" watchObservedRunningTime="2025-10-14 07:55:18.724449466 +0000 UTC m=+251.705688928" Oct 14 07:55:19 crc kubenswrapper[5002]: I1014 07:55:19.711763 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dzd5b" event={"ID":"1cb8dd4f-12fd-46d1-99d3-61b66d6147a1","Type":"ContainerStarted","Data":"10a109498cef89a0e9f64dbee07ab654b6fe8a504b2a2d80d05e2652af77851c"} Oct 14 07:55:19 crc kubenswrapper[5002]: I1014 07:55:19.733364 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dzd5b" podStartSLOduration=3.218983422 podStartE2EDuration="5.733346183s" podCreationTimestamp="2025-10-14 07:55:14 +0000 UTC" firstStartedPulling="2025-10-14 07:55:16.675287366 +0000 UTC m=+249.656526818" lastFinishedPulling="2025-10-14 07:55:19.189650127 +0000 UTC m=+252.170889579" observedRunningTime="2025-10-14 07:55:19.731137964 +0000 UTC m=+252.712377426" watchObservedRunningTime="2025-10-14 07:55:19.733346183 +0000 UTC m=+252.714585635" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.478728 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.479079 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.539882 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.673750 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.673797 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.721382 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.770570 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nggtm" Oct 14 07:55:22 crc kubenswrapper[5002]: I1014 07:55:22.784561 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ts5dh" Oct 14 07:55:24 crc kubenswrapper[5002]: I1014 07:55:24.882938 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:24 crc kubenswrapper[5002]: I1014 07:55:24.884007 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:24 crc kubenswrapper[5002]: I1014 07:55:24.945498 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:25 crc kubenswrapper[5002]: I1014 07:55:25.067739 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:25 crc kubenswrapper[5002]: I1014 07:55:25.067783 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:25 crc kubenswrapper[5002]: I1014 07:55:25.115415 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:25 crc kubenswrapper[5002]: I1014 07:55:25.788831 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6wmt" Oct 14 07:55:25 crc kubenswrapper[5002]: I1014 07:55:25.790173 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dzd5b" Oct 14 07:55:35 crc kubenswrapper[5002]: I1014 07:55:35.602168 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" podUID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" containerName="oauth-openshift" containerID="cri-o://b7046fde711390b84f56d9e2f98b3e9f7c2abadb0c5c1ebea346d817d34431dd" gracePeriod=15 Oct 14 07:55:35 crc kubenswrapper[5002]: I1014 07:55:35.804959 5002 generic.go:334] "Generic (PLEG): container finished" podID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" containerID="b7046fde711390b84f56d9e2f98b3e9f7c2abadb0c5c1ebea346d817d34431dd" exitCode=0 Oct 14 07:55:35 crc kubenswrapper[5002]: I1014 07:55:35.805015 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" event={"ID":"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc","Type":"ContainerDied","Data":"b7046fde711390b84f56d9e2f98b3e9f7c2abadb0c5c1ebea346d817d34431dd"} Oct 14 07:55:35 crc kubenswrapper[5002]: I1014 07:55:35.978472 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.007264 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-s4cf6"] Oct 14 07:55:36 crc kubenswrapper[5002]: E1014 07:55:36.007583 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" containerName="oauth-openshift" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.007604 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" containerName="oauth-openshift" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.007763 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" containerName="oauth-openshift" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.008412 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.021493 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-s4cf6"] Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.056890 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.056949 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.056992 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057015 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057055 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057078 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057102 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qw86\" (UniqueName: \"kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057131 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057153 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057173 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057196 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057231 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057263 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057287 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data\") pod \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\" (UID: \"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc\") " Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057426 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057486 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057513 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057539 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057560 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057584 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057628 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54g6h\" (UniqueName: \"kubernetes.io/projected/31e3f626-a197-490b-af2b-4b007b22935a-kube-api-access-54g6h\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057651 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057675 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057699 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-audit-policies\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057726 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057821 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.057991 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31e3f626-a197-490b-af2b-4b007b22935a-audit-dir\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.058112 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.058146 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.058242 5002 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.059146 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.059514 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.059190 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.059617 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.063598 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.064212 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86" (OuterVolumeSpecName: "kube-api-access-4qw86") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "kube-api-access-4qw86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.064361 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.064799 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.065119 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.075066 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.075352 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.075386 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.075663 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" (UID: "c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.158778 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.158880 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.158925 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.158959 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.158997 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159040 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54g6h\" (UniqueName: \"kubernetes.io/projected/31e3f626-a197-490b-af2b-4b007b22935a-kube-api-access-54g6h\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159076 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159110 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159147 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-audit-policies\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159188 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159231 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31e3f626-a197-490b-af2b-4b007b22935a-audit-dir\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159274 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159313 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159366 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159450 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159473 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159494 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159518 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159555 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159577 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qw86\" (UniqueName: \"kubernetes.io/projected/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-kube-api-access-4qw86\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159600 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159620 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159640 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159660 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159680 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159700 5002 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.159719 5002 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.161434 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.161442 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-service-ca\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.161495 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/31e3f626-a197-490b-af2b-4b007b22935a-audit-dir\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.161798 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.163040 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/31e3f626-a197-490b-af2b-4b007b22935a-audit-policies\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.164928 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-login\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.165231 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-error\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.165506 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-session\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.166213 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.166546 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.166961 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.167758 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.167881 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/31e3f626-a197-490b-af2b-4b007b22935a-v4-0-config-system-router-certs\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.176955 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54g6h\" (UniqueName: \"kubernetes.io/projected/31e3f626-a197-490b-af2b-4b007b22935a-kube-api-access-54g6h\") pod \"oauth-openshift-9565f95f5-s4cf6\" (UID: \"31e3f626-a197-490b-af2b-4b007b22935a\") " pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.325866 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.548064 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-9565f95f5-s4cf6"] Oct 14 07:55:36 crc kubenswrapper[5002]: W1014 07:55:36.554478 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31e3f626_a197_490b_af2b_4b007b22935a.slice/crio-dbf5640bf4280184b6e2db26793f8720b6866e569f27afe81985db4375c1a906 WatchSource:0}: Error finding container dbf5640bf4280184b6e2db26793f8720b6866e569f27afe81985db4375c1a906: Status 404 returned error can't find the container with id dbf5640bf4280184b6e2db26793f8720b6866e569f27afe81985db4375c1a906 Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.810545 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" event={"ID":"c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc","Type":"ContainerDied","Data":"a322e84f397cf4ce0576275f3e644ea0b9d273c8c1bbbd2b583a270e135ab3a1"} Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.810570 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-grb97" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.810596 5002 scope.go:117] "RemoveContainer" containerID="b7046fde711390b84f56d9e2f98b3e9f7c2abadb0c5c1ebea346d817d34431dd" Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.813178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" event={"ID":"31e3f626-a197-490b-af2b-4b007b22935a","Type":"ContainerStarted","Data":"dbf5640bf4280184b6e2db26793f8720b6866e569f27afe81985db4375c1a906"} Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.834828 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:55:36 crc kubenswrapper[5002]: I1014 07:55:36.842133 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-grb97"] Oct 14 07:55:37 crc kubenswrapper[5002]: I1014 07:55:37.727741 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc" path="/var/lib/kubelet/pods/c56dbb4d-ee0d-4fa5-b896-0b1b01b4f5cc/volumes" Oct 14 07:55:37 crc kubenswrapper[5002]: I1014 07:55:37.825981 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" event={"ID":"31e3f626-a197-490b-af2b-4b007b22935a","Type":"ContainerStarted","Data":"463a1509c0ad2f698c84f456c6644450b73cb26238edd890e99b0fc18aed148a"} Oct 14 07:55:37 crc kubenswrapper[5002]: I1014 07:55:37.826237 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:37 crc kubenswrapper[5002]: I1014 07:55:37.835090 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" Oct 14 07:55:37 crc kubenswrapper[5002]: I1014 07:55:37.845428 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-9565f95f5-s4cf6" podStartSLOduration=27.845410886 podStartE2EDuration="27.845410886s" podCreationTimestamp="2025-10-14 07:55:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:55:37.844181021 +0000 UTC m=+270.825420493" watchObservedRunningTime="2025-10-14 07:55:37.845410886 +0000 UTC m=+270.826650338" Oct 14 07:56:39 crc kubenswrapper[5002]: I1014 07:56:39.218322 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:56:39 crc kubenswrapper[5002]: I1014 07:56:39.219137 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:57:09 crc kubenswrapper[5002]: I1014 07:57:09.217998 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:57:09 crc kubenswrapper[5002]: I1014 07:57:09.219062 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.218448 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.219023 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.219082 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.220472 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.220572 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf" gracePeriod=600 Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.655739 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf" exitCode=0 Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.655797 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf"} Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.656125 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3"} Oct 14 07:57:39 crc kubenswrapper[5002]: I1014 07:57:39.656154 5002 scope.go:117] "RemoveContainer" containerID="11e6fd5853a216cda990260a55daec292839ebddd4aac8d4ca82dc6d07097d7b" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.228501 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6gqx4"] Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.229739 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.255060 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6gqx4"] Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.431947 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-bound-sa-token\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432196 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-certificates\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432360 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-tls\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432516 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7df13207-e5e2-4137-b1fb-9e632b96f596-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432632 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kspsl\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-kube-api-access-kspsl\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432777 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7df13207-e5e2-4137-b1fb-9e632b96f596-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432824 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-trusted-ca\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.432920 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.466215 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.533929 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kspsl\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-kube-api-access-kspsl\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534028 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7df13207-e5e2-4137-b1fb-9e632b96f596-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534064 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-trusted-ca\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534125 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-bound-sa-token\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534178 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-certificates\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534211 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-tls\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534253 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7df13207-e5e2-4137-b1fb-9e632b96f596-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.534618 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7df13207-e5e2-4137-b1fb-9e632b96f596-ca-trust-extracted\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.536346 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-certificates\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.536668 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7df13207-e5e2-4137-b1fb-9e632b96f596-trusted-ca\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.542214 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7df13207-e5e2-4137-b1fb-9e632b96f596-installation-pull-secrets\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.542287 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-registry-tls\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.554981 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-bound-sa-token\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.560190 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kspsl\" (UniqueName: \"kubernetes.io/projected/7df13207-e5e2-4137-b1fb-9e632b96f596-kube-api-access-kspsl\") pod \"image-registry-66df7c8f76-6gqx4\" (UID: \"7df13207-e5e2-4137-b1fb-9e632b96f596\") " pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:25 crc kubenswrapper[5002]: I1014 07:58:25.849627 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:26 crc kubenswrapper[5002]: I1014 07:58:26.378543 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-6gqx4"] Oct 14 07:58:27 crc kubenswrapper[5002]: I1014 07:58:27.048259 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" event={"ID":"7df13207-e5e2-4137-b1fb-9e632b96f596","Type":"ContainerStarted","Data":"31ed542cc357297f1a13c04ee14fd98f8a295450c6485114349d1a6b047b2d7e"} Oct 14 07:58:27 crc kubenswrapper[5002]: I1014 07:58:27.048720 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" event={"ID":"7df13207-e5e2-4137-b1fb-9e632b96f596","Type":"ContainerStarted","Data":"c6f7be4d090f17c63b0741f8853f76606d3f8bb167618e0d587d4163ee862af8"} Oct 14 07:58:27 crc kubenswrapper[5002]: I1014 07:58:27.048890 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:45 crc kubenswrapper[5002]: I1014 07:58:45.863799 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" Oct 14 07:58:45 crc kubenswrapper[5002]: I1014 07:58:45.894359 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-6gqx4" podStartSLOduration=20.894336543 podStartE2EDuration="20.894336543s" podCreationTimestamp="2025-10-14 07:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:58:27.082256995 +0000 UTC m=+440.063496507" watchObservedRunningTime="2025-10-14 07:58:45.894336543 +0000 UTC m=+458.875576025" Oct 14 07:58:45 crc kubenswrapper[5002]: I1014 07:58:45.947484 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.007524 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" podUID="53e3879b-b0b6-4bfa-876d-765526e2e952" containerName="registry" containerID="cri-o://fd71c15f1884d3e4a94debda691233a4eb17220dd8a3be032426878a9d691656" gracePeriod=30 Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.352985 5002 generic.go:334] "Generic (PLEG): container finished" podID="53e3879b-b0b6-4bfa-876d-765526e2e952" containerID="fd71c15f1884d3e4a94debda691233a4eb17220dd8a3be032426878a9d691656" exitCode=0 Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.353110 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" event={"ID":"53e3879b-b0b6-4bfa-876d-765526e2e952","Type":"ContainerDied","Data":"fd71c15f1884d3e4a94debda691233a4eb17220dd8a3be032426878a9d691656"} Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.465519 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.472944 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.480601 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574030 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574077 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574121 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574181 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574220 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8fqj\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574297 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574337 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets\") pod \"53e3879b-b0b6-4bfa-876d-765526e2e952\" (UID: \"53e3879b-b0b6-4bfa-876d-765526e2e952\") " Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.574615 5002 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.575410 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.575513 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.579410 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.580715 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.581247 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj" (OuterVolumeSpecName: "kube-api-access-k8fqj") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "kube-api-access-k8fqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.587493 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.601516 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "53e3879b-b0b6-4bfa-876d-765526e2e952" (UID: "53e3879b-b0b6-4bfa-876d-765526e2e952"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.675912 5002 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.675951 5002 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.675968 5002 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/53e3879b-b0b6-4bfa-876d-765526e2e952-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.675983 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8fqj\" (UniqueName: \"kubernetes.io/projected/53e3879b-b0b6-4bfa-876d-765526e2e952-kube-api-access-k8fqj\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.675996 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/53e3879b-b0b6-4bfa-876d-765526e2e952-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:11 crc kubenswrapper[5002]: I1014 07:59:11.676011 5002 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/53e3879b-b0b6-4bfa-876d-765526e2e952-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:59:12 crc kubenswrapper[5002]: I1014 07:59:12.364976 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" event={"ID":"53e3879b-b0b6-4bfa-876d-765526e2e952","Type":"ContainerDied","Data":"ab339dabd582727f379d5add539f7c022738a03784e5154c9ddfbd5d3e3ef8c4"} Oct 14 07:59:12 crc kubenswrapper[5002]: I1014 07:59:12.365068 5002 scope.go:117] "RemoveContainer" containerID="fd71c15f1884d3e4a94debda691233a4eb17220dd8a3be032426878a9d691656" Oct 14 07:59:12 crc kubenswrapper[5002]: I1014 07:59:12.366173 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dc64m" Oct 14 07:59:12 crc kubenswrapper[5002]: I1014 07:59:12.397128 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:59:12 crc kubenswrapper[5002]: I1014 07:59:12.404689 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dc64m"] Oct 14 07:59:13 crc kubenswrapper[5002]: I1014 07:59:13.728487 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53e3879b-b0b6-4bfa-876d-765526e2e952" path="/var/lib/kubelet/pods/53e3879b-b0b6-4bfa-876d-765526e2e952/volumes" Oct 14 07:59:39 crc kubenswrapper[5002]: I1014 07:59:39.218010 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:59:39 crc kubenswrapper[5002]: I1014 07:59:39.218787 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.148380 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8"] Oct 14 08:00:00 crc kubenswrapper[5002]: E1014 08:00:00.149380 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e3879b-b0b6-4bfa-876d-765526e2e952" containerName="registry" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.149402 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e3879b-b0b6-4bfa-876d-765526e2e952" containerName="registry" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.149593 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e3879b-b0b6-4bfa-876d-765526e2e952" containerName="registry" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.150287 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.154371 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.154642 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.169122 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8"] Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.216781 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.216888 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.216998 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k478k\" (UniqueName: \"kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.318215 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k478k\" (UniqueName: \"kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.318296 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.318360 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.320970 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.328691 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.343596 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k478k\" (UniqueName: \"kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k\") pod \"collect-profiles-29340480-cc5x8\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.488132 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:00 crc kubenswrapper[5002]: I1014 08:00:00.736048 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8"] Oct 14 08:00:01 crc kubenswrapper[5002]: I1014 08:00:01.715367 5002 generic.go:334] "Generic (PLEG): container finished" podID="2a0268a0-bec8-43f1-864d-c54e0151fb34" containerID="df850e6f4af91ab4d897f720c99f81649c55f21fbbbedb1ea6455e36220cc3af" exitCode=0 Oct 14 08:00:01 crc kubenswrapper[5002]: I1014 08:00:01.715807 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" event={"ID":"2a0268a0-bec8-43f1-864d-c54e0151fb34","Type":"ContainerDied","Data":"df850e6f4af91ab4d897f720c99f81649c55f21fbbbedb1ea6455e36220cc3af"} Oct 14 08:00:01 crc kubenswrapper[5002]: I1014 08:00:01.715925 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" event={"ID":"2a0268a0-bec8-43f1-864d-c54e0151fb34","Type":"ContainerStarted","Data":"493faf740fd00821afe267775b8569ff6dd1a2036bbfc9b03bdff6d86f979bac"} Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.010374 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.059854 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume\") pod \"2a0268a0-bec8-43f1-864d-c54e0151fb34\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.059938 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k478k\" (UniqueName: \"kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k\") pod \"2a0268a0-bec8-43f1-864d-c54e0151fb34\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.059989 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume\") pod \"2a0268a0-bec8-43f1-864d-c54e0151fb34\" (UID: \"2a0268a0-bec8-43f1-864d-c54e0151fb34\") " Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.060648 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume" (OuterVolumeSpecName: "config-volume") pod "2a0268a0-bec8-43f1-864d-c54e0151fb34" (UID: "2a0268a0-bec8-43f1-864d-c54e0151fb34"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.067934 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2a0268a0-bec8-43f1-864d-c54e0151fb34" (UID: "2a0268a0-bec8-43f1-864d-c54e0151fb34"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.068055 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k" (OuterVolumeSpecName: "kube-api-access-k478k") pod "2a0268a0-bec8-43f1-864d-c54e0151fb34" (UID: "2a0268a0-bec8-43f1-864d-c54e0151fb34"). InnerVolumeSpecName "kube-api-access-k478k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.160960 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a0268a0-bec8-43f1-864d-c54e0151fb34-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.161296 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k478k\" (UniqueName: \"kubernetes.io/projected/2a0268a0-bec8-43f1-864d-c54e0151fb34-kube-api-access-k478k\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.161313 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2a0268a0-bec8-43f1-864d-c54e0151fb34-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.734322 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.736752 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8" event={"ID":"2a0268a0-bec8-43f1-864d-c54e0151fb34","Type":"ContainerDied","Data":"493faf740fd00821afe267775b8569ff6dd1a2036bbfc9b03bdff6d86f979bac"} Oct 14 08:00:03 crc kubenswrapper[5002]: I1014 08:00:03.736810 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="493faf740fd00821afe267775b8569ff6dd1a2036bbfc9b03bdff6d86f979bac" Oct 14 08:00:09 crc kubenswrapper[5002]: I1014 08:00:09.218472 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:00:09 crc kubenswrapper[5002]: I1014 08:00:09.218807 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.425274 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wkw2k"] Oct 14 08:00:30 crc kubenswrapper[5002]: E1014 08:00:30.426006 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0268a0-bec8-43f1-864d-c54e0151fb34" containerName="collect-profiles" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.426023 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0268a0-bec8-43f1-864d-c54e0151fb34" containerName="collect-profiles" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.426153 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0268a0-bec8-43f1-864d-c54e0151fb34" containerName="collect-profiles" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.426526 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.431167 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.431486 5002 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pz5wx" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.431545 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.437861 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hsnk9"] Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.438613 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hsnk9" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.442789 5002 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-ls2dp" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.446966 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wkw2k"] Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.451714 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-66k4z"] Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.452542 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.460282 5002 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-27scs" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.471424 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hsnk9"] Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.475119 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-66k4z"] Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.538396 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bg2\" (UniqueName: \"kubernetes.io/projected/632fb195-4a92-40b0-96e7-4db2227146d5-kube-api-access-g5bg2\") pod \"cert-manager-cainjector-7f985d654d-wkw2k\" (UID: \"632fb195-4a92-40b0-96e7-4db2227146d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.538471 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ghfw\" (UniqueName: \"kubernetes.io/projected/067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7-kube-api-access-8ghfw\") pod \"cert-manager-webhook-5655c58dd6-66k4z\" (UID: \"067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.538500 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqmhl\" (UniqueName: \"kubernetes.io/projected/cbaa9318-0355-41df-b7bd-bd2fe2ce7d30-kube-api-access-cqmhl\") pod \"cert-manager-5b446d88c5-hsnk9\" (UID: \"cbaa9318-0355-41df-b7bd-bd2fe2ce7d30\") " pod="cert-manager/cert-manager-5b446d88c5-hsnk9" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.639568 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ghfw\" (UniqueName: \"kubernetes.io/projected/067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7-kube-api-access-8ghfw\") pod \"cert-manager-webhook-5655c58dd6-66k4z\" (UID: \"067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.639619 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqmhl\" (UniqueName: \"kubernetes.io/projected/cbaa9318-0355-41df-b7bd-bd2fe2ce7d30-kube-api-access-cqmhl\") pod \"cert-manager-5b446d88c5-hsnk9\" (UID: \"cbaa9318-0355-41df-b7bd-bd2fe2ce7d30\") " pod="cert-manager/cert-manager-5b446d88c5-hsnk9" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.639660 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bg2\" (UniqueName: \"kubernetes.io/projected/632fb195-4a92-40b0-96e7-4db2227146d5-kube-api-access-g5bg2\") pod \"cert-manager-cainjector-7f985d654d-wkw2k\" (UID: \"632fb195-4a92-40b0-96e7-4db2227146d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.658667 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ghfw\" (UniqueName: \"kubernetes.io/projected/067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7-kube-api-access-8ghfw\") pod \"cert-manager-webhook-5655c58dd6-66k4z\" (UID: \"067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.659244 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bg2\" (UniqueName: \"kubernetes.io/projected/632fb195-4a92-40b0-96e7-4db2227146d5-kube-api-access-g5bg2\") pod \"cert-manager-cainjector-7f985d654d-wkw2k\" (UID: \"632fb195-4a92-40b0-96e7-4db2227146d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.659785 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqmhl\" (UniqueName: \"kubernetes.io/projected/cbaa9318-0355-41df-b7bd-bd2fe2ce7d30-kube-api-access-cqmhl\") pod \"cert-manager-5b446d88c5-hsnk9\" (UID: \"cbaa9318-0355-41df-b7bd-bd2fe2ce7d30\") " pod="cert-manager/cert-manager-5b446d88c5-hsnk9" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.754082 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.770862 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hsnk9" Oct 14 08:00:30 crc kubenswrapper[5002]: I1014 08:00:30.782815 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.056958 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-66k4z"] Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.067551 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.201554 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hsnk9"] Oct 14 08:00:31 crc kubenswrapper[5002]: W1014 08:00:31.207411 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbaa9318_0355_41df_b7bd_bd2fe2ce7d30.slice/crio-378a9cd3947e0e1b7bc9d0a5ddfbb6c634428545df0a1341027224274010b98e WatchSource:0}: Error finding container 378a9cd3947e0e1b7bc9d0a5ddfbb6c634428545df0a1341027224274010b98e: Status 404 returned error can't find the container with id 378a9cd3947e0e1b7bc9d0a5ddfbb6c634428545df0a1341027224274010b98e Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.213895 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-wkw2k"] Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.928322 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" event={"ID":"632fb195-4a92-40b0-96e7-4db2227146d5","Type":"ContainerStarted","Data":"79f3eb9ce405303650e1533d74bd3db62ee510d953086f6e07b0a269204bb41e"} Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.930629 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" event={"ID":"067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7","Type":"ContainerStarted","Data":"62ba894d77a2cdf5f63334799861facb1c12fa3cbdf8ffd23d92b10f1b0abff0"} Oct 14 08:00:31 crc kubenswrapper[5002]: I1014 08:00:31.932998 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hsnk9" event={"ID":"cbaa9318-0355-41df-b7bd-bd2fe2ce7d30","Type":"ContainerStarted","Data":"378a9cd3947e0e1b7bc9d0a5ddfbb6c634428545df0a1341027224274010b98e"} Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.948488 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" event={"ID":"632fb195-4a92-40b0-96e7-4db2227146d5","Type":"ContainerStarted","Data":"3b28f7177578b977b2405817b171d52bff2eb9c05cbf33504ee7b5d5010fe956"} Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.950788 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" event={"ID":"067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7","Type":"ContainerStarted","Data":"6f9e2febe62fac7fbc10069f857868f3514d47d5390ead6086f0f40730868669"} Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.950972 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.953114 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hsnk9" event={"ID":"cbaa9318-0355-41df-b7bd-bd2fe2ce7d30","Type":"ContainerStarted","Data":"43948d3a281fe7505a25cc1b08439dfb035af2cbc49fdd2a4e82b34f733cb689"} Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.965808 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-wkw2k" podStartSLOduration=2.192861732 podStartE2EDuration="4.965789371s" podCreationTimestamp="2025-10-14 08:00:30 +0000 UTC" firstStartedPulling="2025-10-14 08:00:31.222882345 +0000 UTC m=+564.204121817" lastFinishedPulling="2025-10-14 08:00:33.995810004 +0000 UTC m=+566.977049456" observedRunningTime="2025-10-14 08:00:34.965431533 +0000 UTC m=+567.946671015" watchObservedRunningTime="2025-10-14 08:00:34.965789371 +0000 UTC m=+567.947028823" Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.980336 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" podStartSLOduration=1.993245575 podStartE2EDuration="4.980310592s" podCreationTimestamp="2025-10-14 08:00:30 +0000 UTC" firstStartedPulling="2025-10-14 08:00:31.067341785 +0000 UTC m=+564.048581237" lastFinishedPulling="2025-10-14 08:00:34.054406802 +0000 UTC m=+567.035646254" observedRunningTime="2025-10-14 08:00:34.980309622 +0000 UTC m=+567.961549094" watchObservedRunningTime="2025-10-14 08:00:34.980310592 +0000 UTC m=+567.961550054" Oct 14 08:00:34 crc kubenswrapper[5002]: I1014 08:00:34.994136 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-hsnk9" podStartSLOduration=1.510893958 podStartE2EDuration="4.994116689s" podCreationTimestamp="2025-10-14 08:00:30 +0000 UTC" firstStartedPulling="2025-10-14 08:00:31.210328875 +0000 UTC m=+564.191568337" lastFinishedPulling="2025-10-14 08:00:34.693551626 +0000 UTC m=+567.674791068" observedRunningTime="2025-10-14 08:00:34.99226572 +0000 UTC m=+567.973505192" watchObservedRunningTime="2025-10-14 08:00:34.994116689 +0000 UTC m=+567.975356151" Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.218049 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.218372 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.218423 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.219255 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.219404 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3" gracePeriod=600 Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.990068 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3" exitCode=0 Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.990132 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3"} Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.990565 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f"} Oct 14 08:00:39 crc kubenswrapper[5002]: I1014 08:00:39.990603 5002 scope.go:117] "RemoveContainer" containerID="6460e6b93dcf156fd53e6c3a1eea9a4273f85b9d704a61d119d5695d16116cbf" Oct 14 08:00:40 crc kubenswrapper[5002]: I1014 08:00:40.788577 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-66k4z" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.167147 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gd5gm"] Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.167897 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-controller" containerID="cri-o://d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.167949 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="sbdb" containerID="cri-o://403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.168062 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="nbdb" containerID="cri-o://ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.168136 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="northd" containerID="cri-o://08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.168199 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.168267 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-acl-logging" containerID="cri-o://4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.168361 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-node" containerID="cri-o://187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.210648 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" containerID="cri-o://500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" gracePeriod=30 Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.508879 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/3.log" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.512697 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovn-acl-logging/0.log" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.513253 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovn-controller/0.log" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.513696 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.579804 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kwshp"] Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580053 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kubecfg-setup" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580074 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kubecfg-setup" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580088 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580097 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580108 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580116 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580127 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580135 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580145 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580153 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580162 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="nbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580170 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="nbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580203 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-node" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580211 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-node" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580227 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="northd" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580235 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="northd" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580243 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580250 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580258 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-acl-logging" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580266 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-acl-logging" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580280 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="sbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580287 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="sbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580395 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580408 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580416 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580426 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-acl-logging" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580436 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="nbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580448 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580455 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="kube-rbac-proxy-node" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580464 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="northd" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580474 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovn-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580485 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="sbdb" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580593 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580602 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580707 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: E1014 08:00:41.580821 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580830 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.580973 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerName="ovnkube-controller" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.582614 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627478 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96bqb\" (UniqueName: \"kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627527 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627554 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627574 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627612 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627640 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627687 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627712 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627733 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627754 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627780 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627802 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627859 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627803 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627887 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627919 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627936 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627955 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627952 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.627974 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628004 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628028 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes\") pod \"858331e7-9f04-4a5d-8bf9-8307dfa68556\" (UID: \"858331e7-9f04-4a5d-8bf9-8307dfa68556\") " Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628129 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket" (OuterVolumeSpecName: "log-socket") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628181 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628170 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash" (OuterVolumeSpecName: "host-slash") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628211 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628232 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628236 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628265 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628436 5002 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628450 5002 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-slash\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628459 5002 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628470 5002 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-log-socket\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628481 5002 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628490 5002 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628499 5002 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628509 5002 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628518 5002 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628519 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628551 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log" (OuterVolumeSpecName: "node-log") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628548 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628571 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628593 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628597 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628613 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.628817 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.633370 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.633431 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb" (OuterVolumeSpecName: "kube-api-access-96bqb") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "kube-api-access-96bqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.643052 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "858331e7-9f04-4a5d-8bf9-8307dfa68556" (UID: "858331e7-9f04-4a5d-8bf9-8307dfa68556"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.729769 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-kubelet\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.729825 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-log-socket\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.729894 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.729964 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-netd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.729990 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-systemd-units\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730007 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730023 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-env-overrides\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730046 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-systemd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730064 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-bin\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730177 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-node-log\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730274 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-script-lib\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730349 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56j5l\" (UniqueName: \"kubernetes.io/projected/aa64956f-e101-4543-89ba-210fb3245931-kube-api-access-56j5l\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730382 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-var-lib-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730416 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-etc-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730477 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730510 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-slash\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730542 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-config\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730592 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-netns\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730646 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa64956f-e101-4543-89ba-210fb3245931-ovn-node-metrics-cert\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730688 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-ovn\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730773 5002 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730797 5002 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730817 5002 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730881 5002 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730901 5002 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730921 5002 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730940 5002 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-node-log\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730959 5002 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730980 5002 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/858331e7-9f04-4a5d-8bf9-8307dfa68556-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.730997 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96bqb\" (UniqueName: \"kubernetes.io/projected/858331e7-9f04-4a5d-8bf9-8307dfa68556-kube-api-access-96bqb\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.731015 5002 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/858331e7-9f04-4a5d-8bf9-8307dfa68556-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.832998 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-systemd-units\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833093 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833135 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-env-overrides\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833176 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-systemd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833209 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-bin\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833246 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-node-log\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833294 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-script-lib\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833345 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-var-lib-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833375 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56j5l\" (UniqueName: \"kubernetes.io/projected/aa64956f-e101-4543-89ba-210fb3245931-kube-api-access-56j5l\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833406 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-etc-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833437 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833465 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-slash\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833496 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-config\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833546 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-netns\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.833683 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa64956f-e101-4543-89ba-210fb3245931-ovn-node-metrics-cert\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834112 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-netns\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834146 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-bin\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834259 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-systemd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834313 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-slash\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834374 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-etc-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834370 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-node-log\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834333 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-systemd-units\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834447 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-var-lib-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834437 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-ovn\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.834687 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-ovn\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835448 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-env-overrides\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835504 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-config\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835732 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-kubelet\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-log-socket\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835883 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835934 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aa64956f-e101-4543-89ba-210fb3245931-ovnkube-script-lib\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835964 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-log-socket\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835994 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-netd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835941 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-cni-netd\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835967 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-run-ovn-kubernetes\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835665 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-run-openvswitch\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.835882 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aa64956f-e101-4543-89ba-210fb3245931-host-kubelet\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.840377 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa64956f-e101-4543-89ba-210fb3245931-ovn-node-metrics-cert\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.864540 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56j5l\" (UniqueName: \"kubernetes.io/projected/aa64956f-e101-4543-89ba-210fb3245931-kube-api-access-56j5l\") pod \"ovnkube-node-kwshp\" (UID: \"aa64956f-e101-4543-89ba-210fb3245931\") " pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: I1014 08:00:41.900651 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:41 crc kubenswrapper[5002]: W1014 08:00:41.928228 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa64956f_e101_4543_89ba_210fb3245931.slice/crio-69348bb92c3478c8f11b5378c2b038c392544d5381567686a717964ced1b93f1 WatchSource:0}: Error finding container 69348bb92c3478c8f11b5378c2b038c392544d5381567686a717964ced1b93f1: Status 404 returned error can't find the container with id 69348bb92c3478c8f11b5378c2b038c392544d5381567686a717964ced1b93f1 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.004283 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"69348bb92c3478c8f11b5378c2b038c392544d5381567686a717964ced1b93f1"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.006769 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovnkube-controller/3.log" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.009006 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovn-acl-logging/0.log" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.009518 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-gd5gm_858331e7-9f04-4a5d-8bf9-8307dfa68556/ovn-controller/0.log" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010037 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010059 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010066 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010074 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010080 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010087 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" exitCode=0 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010094 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" exitCode=143 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010100 5002 generic.go:334] "Generic (PLEG): container finished" podID="858331e7-9f04-4a5d-8bf9-8307dfa68556" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" exitCode=143 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010122 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010144 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010189 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010206 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010217 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010227 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010236 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010245 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010266 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010268 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010333 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010344 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010349 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010355 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010360 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010365 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010371 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010386 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010402 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010408 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010414 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010419 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010426 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010430 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010436 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010441 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010446 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010450 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010458 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010465 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010472 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010478 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010483 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010488 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010493 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010498 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010504 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010509 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010514 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010521 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gd5gm" event={"ID":"858331e7-9f04-4a5d-8bf9-8307dfa68556","Type":"ContainerDied","Data":"596453ff2a17a47d84491528bb31fc130ab25c0dbf9ee71db4b931d4af77ffbc"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010528 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010533 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010538 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010543 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010548 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010553 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010558 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010563 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010623 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.010629 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.012413 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/2.log" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.013194 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/1.log" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.013249 5002 generic.go:334] "Generic (PLEG): container finished" podID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" containerID="936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a" exitCode=2 Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.013289 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerDied","Data":"936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.013319 5002 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64"} Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.013952 5002 scope.go:117] "RemoveContainer" containerID="936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.014123 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9n4rz_openshift-multus(359daa94-9198-48cf-bbea-a7d8cdb8f3c7)\"" pod="openshift-multus/multus-9n4rz" podUID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.040858 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.044458 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gd5gm"] Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.055273 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gd5gm"] Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.082392 5002 scope.go:117] "RemoveContainer" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.095744 5002 scope.go:117] "RemoveContainer" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.108991 5002 scope.go:117] "RemoveContainer" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.120964 5002 scope.go:117] "RemoveContainer" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.131606 5002 scope.go:117] "RemoveContainer" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.143520 5002 scope.go:117] "RemoveContainer" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.162359 5002 scope.go:117] "RemoveContainer" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.181694 5002 scope.go:117] "RemoveContainer" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.195918 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.196383 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.196444 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} err="failed to get container status \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.196478 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.196812 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": container with ID starting with c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e not found: ID does not exist" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.196874 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} err="failed to get container status \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": rpc error: code = NotFound desc = could not find container \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": container with ID starting with c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.196900 5002 scope.go:117] "RemoveContainer" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.197152 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": container with ID starting with 403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705 not found: ID does not exist" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197184 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} err="failed to get container status \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": rpc error: code = NotFound desc = could not find container \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": container with ID starting with 403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197206 5002 scope.go:117] "RemoveContainer" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.197437 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": container with ID starting with ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d not found: ID does not exist" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197463 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} err="failed to get container status \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": rpc error: code = NotFound desc = could not find container \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": container with ID starting with ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197478 5002 scope.go:117] "RemoveContainer" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.197739 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": container with ID starting with 08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce not found: ID does not exist" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197783 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} err="failed to get container status \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": rpc error: code = NotFound desc = could not find container \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": container with ID starting with 08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.197816 5002 scope.go:117] "RemoveContainer" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.198143 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": container with ID starting with 973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae not found: ID does not exist" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198173 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} err="failed to get container status \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": rpc error: code = NotFound desc = could not find container \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": container with ID starting with 973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198192 5002 scope.go:117] "RemoveContainer" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.198419 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": container with ID starting with 187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db not found: ID does not exist" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198447 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} err="failed to get container status \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": rpc error: code = NotFound desc = could not find container \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": container with ID starting with 187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198466 5002 scope.go:117] "RemoveContainer" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.198756 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": container with ID starting with 4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e not found: ID does not exist" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198790 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} err="failed to get container status \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": rpc error: code = NotFound desc = could not find container \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": container with ID starting with 4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.198812 5002 scope.go:117] "RemoveContainer" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.199247 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": container with ID starting with d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8 not found: ID does not exist" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.199277 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} err="failed to get container status \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": rpc error: code = NotFound desc = could not find container \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": container with ID starting with d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.199295 5002 scope.go:117] "RemoveContainer" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: E1014 08:00:42.199649 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": container with ID starting with 3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed not found: ID does not exist" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.199672 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} err="failed to get container status \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": rpc error: code = NotFound desc = could not find container \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": container with ID starting with 3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.199690 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200014 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} err="failed to get container status \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200048 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200337 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} err="failed to get container status \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": rpc error: code = NotFound desc = could not find container \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": container with ID starting with c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200374 5002 scope.go:117] "RemoveContainer" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200699 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} err="failed to get container status \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": rpc error: code = NotFound desc = could not find container \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": container with ID starting with 403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.200724 5002 scope.go:117] "RemoveContainer" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201014 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} err="failed to get container status \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": rpc error: code = NotFound desc = could not find container \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": container with ID starting with ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201043 5002 scope.go:117] "RemoveContainer" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201305 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} err="failed to get container status \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": rpc error: code = NotFound desc = could not find container \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": container with ID starting with 08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201375 5002 scope.go:117] "RemoveContainer" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201814 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} err="failed to get container status \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": rpc error: code = NotFound desc = could not find container \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": container with ID starting with 973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.201859 5002 scope.go:117] "RemoveContainer" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.202176 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} err="failed to get container status \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": rpc error: code = NotFound desc = could not find container \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": container with ID starting with 187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.202203 5002 scope.go:117] "RemoveContainer" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.202785 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} err="failed to get container status \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": rpc error: code = NotFound desc = could not find container \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": container with ID starting with 4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.202808 5002 scope.go:117] "RemoveContainer" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.203208 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} err="failed to get container status \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": rpc error: code = NotFound desc = could not find container \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": container with ID starting with d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.203273 5002 scope.go:117] "RemoveContainer" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.203695 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} err="failed to get container status \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": rpc error: code = NotFound desc = could not find container \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": container with ID starting with 3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.203718 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204056 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} err="failed to get container status \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204081 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204311 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} err="failed to get container status \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": rpc error: code = NotFound desc = could not find container \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": container with ID starting with c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204336 5002 scope.go:117] "RemoveContainer" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204664 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} err="failed to get container status \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": rpc error: code = NotFound desc = could not find container \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": container with ID starting with 403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.204685 5002 scope.go:117] "RemoveContainer" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205014 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} err="failed to get container status \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": rpc error: code = NotFound desc = could not find container \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": container with ID starting with ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205045 5002 scope.go:117] "RemoveContainer" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205315 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} err="failed to get container status \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": rpc error: code = NotFound desc = could not find container \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": container with ID starting with 08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205356 5002 scope.go:117] "RemoveContainer" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205611 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} err="failed to get container status \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": rpc error: code = NotFound desc = could not find container \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": container with ID starting with 973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205636 5002 scope.go:117] "RemoveContainer" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.205993 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} err="failed to get container status \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": rpc error: code = NotFound desc = could not find container \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": container with ID starting with 187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.206019 5002 scope.go:117] "RemoveContainer" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.206344 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} err="failed to get container status \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": rpc error: code = NotFound desc = could not find container \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": container with ID starting with 4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.206368 5002 scope.go:117] "RemoveContainer" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.206641 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} err="failed to get container status \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": rpc error: code = NotFound desc = could not find container \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": container with ID starting with d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.206703 5002 scope.go:117] "RemoveContainer" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207080 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} err="failed to get container status \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": rpc error: code = NotFound desc = could not find container \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": container with ID starting with 3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207109 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207356 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} err="failed to get container status \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207389 5002 scope.go:117] "RemoveContainer" containerID="c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207654 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e"} err="failed to get container status \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": rpc error: code = NotFound desc = could not find container \"c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e\": container with ID starting with c2fc60e8a557dc16763dfcb370612e7c0f95011a6681cf93783c9369d417667e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207704 5002 scope.go:117] "RemoveContainer" containerID="403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.207997 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705"} err="failed to get container status \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": rpc error: code = NotFound desc = could not find container \"403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705\": container with ID starting with 403ffd68d8b5fa3c70da9d1e8a59ed3425490087087040ad59e6ab1452868705 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208025 5002 scope.go:117] "RemoveContainer" containerID="ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208247 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d"} err="failed to get container status \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": rpc error: code = NotFound desc = could not find container \"ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d\": container with ID starting with ec49952023e1427e39f90043e82df9fda62db8be4576dc93ec05ecaef990693d not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208272 5002 scope.go:117] "RemoveContainer" containerID="08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208510 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce"} err="failed to get container status \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": rpc error: code = NotFound desc = could not find container \"08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce\": container with ID starting with 08b5eadbb5383fe8b984826deb7d74582acfaf0f74ee15d88559346a89dab2ce not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208546 5002 scope.go:117] "RemoveContainer" containerID="973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208777 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae"} err="failed to get container status \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": rpc error: code = NotFound desc = could not find container \"973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae\": container with ID starting with 973f85604091a6afbfa4cb6b04d13b4b50052bcfd1f01b8816ab8ec84ca8edae not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.208800 5002 scope.go:117] "RemoveContainer" containerID="187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209109 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db"} err="failed to get container status \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": rpc error: code = NotFound desc = could not find container \"187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db\": container with ID starting with 187ec63075b70ac2c50c09861859d5a0cee8e3507801c101ccb5a9f74f8dd8db not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209140 5002 scope.go:117] "RemoveContainer" containerID="4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209438 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e"} err="failed to get container status \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": rpc error: code = NotFound desc = could not find container \"4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e\": container with ID starting with 4287d6155be31a735087e5cd6f33848099e575622859f7e1c286d4731ce1a86e not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209475 5002 scope.go:117] "RemoveContainer" containerID="d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209762 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8"} err="failed to get container status \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": rpc error: code = NotFound desc = could not find container \"d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8\": container with ID starting with d1fb7da6a3a42432ab9eafe65ea7a423d81e26fb9af41e986c394a67431619f8 not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.209813 5002 scope.go:117] "RemoveContainer" containerID="3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.210235 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed"} err="failed to get container status \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": rpc error: code = NotFound desc = could not find container \"3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed\": container with ID starting with 3d43bb212ff03270c8ae5a2ecbd50b8f94f64b819fdb059564c0d0cb07f2b8ed not found: ID does not exist" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.210263 5002 scope.go:117] "RemoveContainer" containerID="500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d" Oct 14 08:00:42 crc kubenswrapper[5002]: I1014 08:00:42.210500 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d"} err="failed to get container status \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": rpc error: code = NotFound desc = could not find container \"500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d\": container with ID starting with 500dfe6c8fb7bc2cb3bbd0a8714fb6c9295aa269b81a9cffcdba1f9e2c1fce4d not found: ID does not exist" Oct 14 08:00:43 crc kubenswrapper[5002]: I1014 08:00:43.022922 5002 generic.go:334] "Generic (PLEG): container finished" podID="aa64956f-e101-4543-89ba-210fb3245931" containerID="e8c31b422d004cedd422ae886c0f147c6c7b538a215a0e39d2cad77360ddf8d2" exitCode=0 Oct 14 08:00:43 crc kubenswrapper[5002]: I1014 08:00:43.022999 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerDied","Data":"e8c31b422d004cedd422ae886c0f147c6c7b538a215a0e39d2cad77360ddf8d2"} Oct 14 08:00:43 crc kubenswrapper[5002]: I1014 08:00:43.728712 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="858331e7-9f04-4a5d-8bf9-8307dfa68556" path="/var/lib/kubelet/pods/858331e7-9f04-4a5d-8bf9-8307dfa68556/volumes" Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032504 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"3e1bb8fb494b4fdd5cf528838edab900493db192d02adbbd4112a90b5c875ef9"} Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032558 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"74a578ade9378662125c546d38e22f0641630f593aa1cf2f34dafa5f5a3bc522"} Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032571 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"d7753617798d4b2d1f61f992868a79bc1c0586b44a5e453027d09bcc4c7b06b3"} Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032582 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"6755598ac4ac6547dbcf1a21a2d6bf830f6dcd8e7747144c4409803e9ec27a76"} Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032592 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"a8988eeee70ff7823d5a7bc2dc9bcd6c3d8720dc0581f87f1b906594bb7e75f9"} Oct 14 08:00:44 crc kubenswrapper[5002]: I1014 08:00:44.032605 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"ec33d0b4b632c4fafbb0cdecc3adb96079fbac5a32c578f14ee6a66eba252833"} Oct 14 08:00:47 crc kubenswrapper[5002]: I1014 08:00:47.062691 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"5bcc440e2b86590258bb35bd294e5c31026c781c709ef23735b5bce74b22f51e"} Oct 14 08:00:49 crc kubenswrapper[5002]: I1014 08:00:49.077216 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" event={"ID":"aa64956f-e101-4543-89ba-210fb3245931","Type":"ContainerStarted","Data":"75bd2d00ff4ed2b80df8fd2b21e0a64c2e3d1504b93123b9c0e82955646e4c18"} Oct 14 08:00:49 crc kubenswrapper[5002]: I1014 08:00:49.077598 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:49 crc kubenswrapper[5002]: I1014 08:00:49.077634 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:49 crc kubenswrapper[5002]: I1014 08:00:49.102792 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:49 crc kubenswrapper[5002]: I1014 08:00:49.109031 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" podStartSLOduration=8.109013463 podStartE2EDuration="8.109013463s" podCreationTimestamp="2025-10-14 08:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:00:49.104879672 +0000 UTC m=+582.086119154" watchObservedRunningTime="2025-10-14 08:00:49.109013463 +0000 UTC m=+582.090252935" Oct 14 08:00:50 crc kubenswrapper[5002]: I1014 08:00:50.084253 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:50 crc kubenswrapper[5002]: I1014 08:00:50.192512 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:00:57 crc kubenswrapper[5002]: I1014 08:00:57.726796 5002 scope.go:117] "RemoveContainer" containerID="936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a" Oct 14 08:00:57 crc kubenswrapper[5002]: E1014 08:00:57.728022 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-9n4rz_openshift-multus(359daa94-9198-48cf-bbea-a7d8cdb8f3c7)\"" pod="openshift-multus/multus-9n4rz" podUID="359daa94-9198-48cf-bbea-a7d8cdb8f3c7" Oct 14 08:01:07 crc kubenswrapper[5002]: I1014 08:01:07.941476 5002 scope.go:117] "RemoveContainer" containerID="35aa50b87c81cc4a882665ae2cd3eadc4f6f1be3a19b379bf29f6b6a83c86c64" Oct 14 08:01:08 crc kubenswrapper[5002]: I1014 08:01:08.203439 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/2.log" Oct 14 08:01:11 crc kubenswrapper[5002]: I1014 08:01:11.720886 5002 scope.go:117] "RemoveContainer" containerID="936b3930a13a0fe1bf16ab64d31d307c08e2dafe2fad10715756968e3ae7d95a" Oct 14 08:01:11 crc kubenswrapper[5002]: I1014 08:01:11.932831 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kwshp" Oct 14 08:01:12 crc kubenswrapper[5002]: I1014 08:01:12.230250 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9n4rz_359daa94-9198-48cf-bbea-a7d8cdb8f3c7/kube-multus/2.log" Oct 14 08:01:12 crc kubenswrapper[5002]: I1014 08:01:12.230351 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9n4rz" event={"ID":"359daa94-9198-48cf-bbea-a7d8cdb8f3c7","Type":"ContainerStarted","Data":"0fffb4dda9dae1cc71875586cc6c71ae62573b86c9ba4f58c4ca33af6707fa26"} Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.674101 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv"] Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.676689 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.679692 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.691958 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv"] Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.766387 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.766483 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwqt6\" (UniqueName: \"kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.766533 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.867973 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.868076 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwqt6\" (UniqueName: \"kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.868135 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.869014 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.869107 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:26 crc kubenswrapper[5002]: I1014 08:01:26.900916 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwqt6\" (UniqueName: \"kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:27 crc kubenswrapper[5002]: I1014 08:01:27.011188 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:27 crc kubenswrapper[5002]: I1014 08:01:27.278827 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv"] Oct 14 08:01:27 crc kubenswrapper[5002]: I1014 08:01:27.316468 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" event={"ID":"909ae76b-abca-406b-9389-49b934d5f1fc","Type":"ContainerStarted","Data":"19fd8ade670d4f5b245ccb8d29b34a33c23ad64eae8c5f38e94a3f2134a6f115"} Oct 14 08:01:28 crc kubenswrapper[5002]: I1014 08:01:28.326043 5002 generic.go:334] "Generic (PLEG): container finished" podID="909ae76b-abca-406b-9389-49b934d5f1fc" containerID="10d01bb1521c3d71e8aabfff467d258521c8f5466a0d58cf4ccc654e44b08cea" exitCode=0 Oct 14 08:01:28 crc kubenswrapper[5002]: I1014 08:01:28.326122 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" event={"ID":"909ae76b-abca-406b-9389-49b934d5f1fc","Type":"ContainerDied","Data":"10d01bb1521c3d71e8aabfff467d258521c8f5466a0d58cf4ccc654e44b08cea"} Oct 14 08:01:30 crc kubenswrapper[5002]: I1014 08:01:30.346518 5002 generic.go:334] "Generic (PLEG): container finished" podID="909ae76b-abca-406b-9389-49b934d5f1fc" containerID="ca9885942de03b29035cdca7da8dce6e1a89b39efbe110fd91a3fd2d2da6fd92" exitCode=0 Oct 14 08:01:30 crc kubenswrapper[5002]: I1014 08:01:30.346649 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" event={"ID":"909ae76b-abca-406b-9389-49b934d5f1fc","Type":"ContainerDied","Data":"ca9885942de03b29035cdca7da8dce6e1a89b39efbe110fd91a3fd2d2da6fd92"} Oct 14 08:01:31 crc kubenswrapper[5002]: I1014 08:01:31.357131 5002 generic.go:334] "Generic (PLEG): container finished" podID="909ae76b-abca-406b-9389-49b934d5f1fc" containerID="54aac28c38b7d9e5b31834ca809deeaf8c4910dae6c85447827102610cdf12b7" exitCode=0 Oct 14 08:01:31 crc kubenswrapper[5002]: I1014 08:01:31.357191 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" event={"ID":"909ae76b-abca-406b-9389-49b934d5f1fc","Type":"ContainerDied","Data":"54aac28c38b7d9e5b31834ca809deeaf8c4910dae6c85447827102610cdf12b7"} Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.655659 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.749983 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwqt6\" (UniqueName: \"kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6\") pod \"909ae76b-abca-406b-9389-49b934d5f1fc\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.750152 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util\") pod \"909ae76b-abca-406b-9389-49b934d5f1fc\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.750261 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle\") pod \"909ae76b-abca-406b-9389-49b934d5f1fc\" (UID: \"909ae76b-abca-406b-9389-49b934d5f1fc\") " Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.750828 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle" (OuterVolumeSpecName: "bundle") pod "909ae76b-abca-406b-9389-49b934d5f1fc" (UID: "909ae76b-abca-406b-9389-49b934d5f1fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.756531 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6" (OuterVolumeSpecName: "kube-api-access-dwqt6") pod "909ae76b-abca-406b-9389-49b934d5f1fc" (UID: "909ae76b-abca-406b-9389-49b934d5f1fc"). InnerVolumeSpecName "kube-api-access-dwqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.852346 5002 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.852389 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwqt6\" (UniqueName: \"kubernetes.io/projected/909ae76b-abca-406b-9389-49b934d5f1fc-kube-api-access-dwqt6\") on node \"crc\" DevicePath \"\"" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.899022 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util" (OuterVolumeSpecName: "util") pod "909ae76b-abca-406b-9389-49b934d5f1fc" (UID: "909ae76b-abca-406b-9389-49b934d5f1fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:01:32 crc kubenswrapper[5002]: I1014 08:01:32.953602 5002 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/909ae76b-abca-406b-9389-49b934d5f1fc-util\") on node \"crc\" DevicePath \"\"" Oct 14 08:01:33 crc kubenswrapper[5002]: I1014 08:01:33.373887 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" event={"ID":"909ae76b-abca-406b-9389-49b934d5f1fc","Type":"ContainerDied","Data":"19fd8ade670d4f5b245ccb8d29b34a33c23ad64eae8c5f38e94a3f2134a6f115"} Oct 14 08:01:33 crc kubenswrapper[5002]: I1014 08:01:33.373956 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19fd8ade670d4f5b245ccb8d29b34a33c23ad64eae8c5f38e94a3f2134a6f115" Oct 14 08:01:33 crc kubenswrapper[5002]: I1014 08:01:33.373973 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.247012 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt"] Oct 14 08:01:38 crc kubenswrapper[5002]: E1014 08:01:38.248175 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="util" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.248203 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="util" Oct 14 08:01:38 crc kubenswrapper[5002]: E1014 08:01:38.248215 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="pull" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.248222 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="pull" Oct 14 08:01:38 crc kubenswrapper[5002]: E1014 08:01:38.248231 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="extract" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.248238 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="extract" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.248360 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="909ae76b-abca-406b-9389-49b934d5f1fc" containerName="extract" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.248755 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.251557 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bwlxl" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.252167 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.253468 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.260615 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt"] Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.331458 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fxz8\" (UniqueName: \"kubernetes.io/projected/38ea0de3-d753-486f-8fc0-b5239e804cb2-kube-api-access-2fxz8\") pod \"nmstate-operator-858ddd8f98-hpwzt\" (UID: \"38ea0de3-d753-486f-8fc0-b5239e804cb2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.432506 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fxz8\" (UniqueName: \"kubernetes.io/projected/38ea0de3-d753-486f-8fc0-b5239e804cb2-kube-api-access-2fxz8\") pod \"nmstate-operator-858ddd8f98-hpwzt\" (UID: \"38ea0de3-d753-486f-8fc0-b5239e804cb2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.456255 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fxz8\" (UniqueName: \"kubernetes.io/projected/38ea0de3-d753-486f-8fc0-b5239e804cb2-kube-api-access-2fxz8\") pod \"nmstate-operator-858ddd8f98-hpwzt\" (UID: \"38ea0de3-d753-486f-8fc0-b5239e804cb2\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.574284 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" Oct 14 08:01:38 crc kubenswrapper[5002]: I1014 08:01:38.779079 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt"] Oct 14 08:01:38 crc kubenswrapper[5002]: W1014 08:01:38.790229 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38ea0de3_d753_486f_8fc0_b5239e804cb2.slice/crio-bc09a648e6b5c47944b94a76fc6a89c25e15f143c3fc35f7a77f48641d90128b WatchSource:0}: Error finding container bc09a648e6b5c47944b94a76fc6a89c25e15f143c3fc35f7a77f48641d90128b: Status 404 returned error can't find the container with id bc09a648e6b5c47944b94a76fc6a89c25e15f143c3fc35f7a77f48641d90128b Oct 14 08:01:39 crc kubenswrapper[5002]: I1014 08:01:39.412949 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" event={"ID":"38ea0de3-d753-486f-8fc0-b5239e804cb2","Type":"ContainerStarted","Data":"bc09a648e6b5c47944b94a76fc6a89c25e15f143c3fc35f7a77f48641d90128b"} Oct 14 08:01:41 crc kubenswrapper[5002]: I1014 08:01:41.427882 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" event={"ID":"38ea0de3-d753-486f-8fc0-b5239e804cb2","Type":"ContainerStarted","Data":"ef3004e9bbc3e8b5f00831ef80db06bd671464cdaf48c1c0159b0d9848ba4860"} Oct 14 08:01:41 crc kubenswrapper[5002]: I1014 08:01:41.450045 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hpwzt" podStartSLOduration=1.261526717 podStartE2EDuration="3.450014532s" podCreationTimestamp="2025-10-14 08:01:38 +0000 UTC" firstStartedPulling="2025-10-14 08:01:38.792516382 +0000 UTC m=+631.773755844" lastFinishedPulling="2025-10-14 08:01:40.981004197 +0000 UTC m=+633.962243659" observedRunningTime="2025-10-14 08:01:41.448919844 +0000 UTC m=+634.430159346" watchObservedRunningTime="2025-10-14 08:01:41.450014532 +0000 UTC m=+634.431254034" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.782889 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.784736 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" Oct 14 08:01:46 crc kubenswrapper[5002]: W1014 08:01:46.786965 5002 reflector.go:561] object-"openshift-nmstate"/"nmstate-handler-dockercfg-qcmhp": failed to list *v1.Secret: secrets "nmstate-handler-dockercfg-qcmhp" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Oct 14 08:01:46 crc kubenswrapper[5002]: E1014 08:01:46.787019 5002 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nmstate-handler-dockercfg-qcmhp\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nmstate-handler-dockercfg-qcmhp\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.797529 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.798928 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.800630 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.801699 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.810452 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-82rj9"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.811706 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.819704 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.899928 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.900518 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.901986 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-fzbfd" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.902068 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.902097 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.939399 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm"] Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960412 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5z4\" (UniqueName: \"kubernetes.io/projected/cb6927e5-34dd-43d1-9dea-db00339df27e-kube-api-access-9p5z4\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960489 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp7fr\" (UniqueName: \"kubernetes.io/projected/89f2e467-845f-45fb-b411-48c0cf2d3f0d-kube-api-access-pp7fr\") pod \"nmstate-metrics-fdff9cb8d-dq7k8\" (UID: \"89f2e467-845f-45fb-b411-48c0cf2d3f0d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960530 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vhm9\" (UniqueName: \"kubernetes.io/projected/0629a9d2-3340-4e00-9673-baaaf354a8ce-kube-api-access-7vhm9\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960560 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-nmstate-lock\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960652 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-dbus-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960699 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-ovs-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:46 crc kubenswrapper[5002]: I1014 08:01:46.960723 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb6927e5-34dd-43d1-9dea-db00339df27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.061761 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/59bfeb19-1a15-4cbd-8970-5c18ff149992-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.061890 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp7fr\" (UniqueName: \"kubernetes.io/projected/89f2e467-845f-45fb-b411-48c0cf2d3f0d-kube-api-access-pp7fr\") pod \"nmstate-metrics-fdff9cb8d-dq7k8\" (UID: \"89f2e467-845f-45fb-b411-48c0cf2d3f0d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062276 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vhm9\" (UniqueName: \"kubernetes.io/projected/0629a9d2-3340-4e00-9673-baaaf354a8ce-kube-api-access-7vhm9\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062323 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062356 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-nmstate-lock\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062413 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vnv9\" (UniqueName: \"kubernetes.io/projected/59bfeb19-1a15-4cbd-8970-5c18ff149992-kube-api-access-2vnv9\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-nmstate-lock\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062466 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-dbus-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062495 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb6927e5-34dd-43d1-9dea-db00339df27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062509 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-ovs-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062530 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5z4\" (UniqueName: \"kubernetes.io/projected/cb6927e5-34dd-43d1-9dea-db00339df27e-kube-api-access-9p5z4\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062680 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-dbus-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.062713 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/0629a9d2-3340-4e00-9673-baaaf354a8ce-ovs-socket\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.081477 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb6927e5-34dd-43d1-9dea-db00339df27e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.084625 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vhm9\" (UniqueName: \"kubernetes.io/projected/0629a9d2-3340-4e00-9673-baaaf354a8ce-kube-api-access-7vhm9\") pod \"nmstate-handler-82rj9\" (UID: \"0629a9d2-3340-4e00-9673-baaaf354a8ce\") " pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.085899 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5z4\" (UniqueName: \"kubernetes.io/projected/cb6927e5-34dd-43d1-9dea-db00339df27e-kube-api-access-9p5z4\") pod \"nmstate-webhook-6cdbc54649-4tmr2\" (UID: \"cb6927e5-34dd-43d1-9dea-db00339df27e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.090556 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp7fr\" (UniqueName: \"kubernetes.io/projected/89f2e467-845f-45fb-b411-48c0cf2d3f0d-kube-api-access-pp7fr\") pod \"nmstate-metrics-fdff9cb8d-dq7k8\" (UID: \"89f2e467-845f-45fb-b411-48c0cf2d3f0d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.142104 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-69797c4689-gjj84"] Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.142695 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.154109 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69797c4689-gjj84"] Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.163403 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vnv9\" (UniqueName: \"kubernetes.io/projected/59bfeb19-1a15-4cbd-8970-5c18ff149992-kube-api-access-2vnv9\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.163514 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/59bfeb19-1a15-4cbd-8970-5c18ff149992-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.163551 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: E1014 08:01:47.163653 5002 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 14 08:01:47 crc kubenswrapper[5002]: E1014 08:01:47.163717 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert podName:59bfeb19-1a15-4cbd-8970-5c18ff149992 nodeName:}" failed. No retries permitted until 2025-10-14 08:01:47.663697358 +0000 UTC m=+640.644936810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-s9gvm" (UID: "59bfeb19-1a15-4cbd-8970-5c18ff149992") : secret "plugin-serving-cert" not found Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.164403 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/59bfeb19-1a15-4cbd-8970-5c18ff149992-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.195140 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vnv9\" (UniqueName: \"kubernetes.io/projected/59bfeb19-1a15-4cbd-8970-5c18ff149992-kube-api-access-2vnv9\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264469 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-trusted-ca-bundle\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264605 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-oauth-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264643 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-service-ca\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264672 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264702 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264726 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-oauth-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.264768 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxh7t\" (UniqueName: \"kubernetes.io/projected/50e01b14-76e1-4d96-bc0e-38e89c3061ac-kube-api-access-wxh7t\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366411 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-service-ca\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366481 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366520 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366548 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-oauth-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366581 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxh7t\" (UniqueName: \"kubernetes.io/projected/50e01b14-76e1-4d96-bc0e-38e89c3061ac-kube-api-access-wxh7t\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366677 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-trusted-ca-bundle\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.366771 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-oauth-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.367822 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-trusted-ca-bundle\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.367963 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-service-ca\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.368264 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.368493 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/50e01b14-76e1-4d96-bc0e-38e89c3061ac-oauth-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.373233 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-serving-cert\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.377996 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/50e01b14-76e1-4d96-bc0e-38e89c3061ac-console-oauth-config\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.402621 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxh7t\" (UniqueName: \"kubernetes.io/projected/50e01b14-76e1-4d96-bc0e-38e89c3061ac-kube-api-access-wxh7t\") pod \"console-69797c4689-gjj84\" (UID: \"50e01b14-76e1-4d96-bc0e-38e89c3061ac\") " pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.455753 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.646796 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-qcmhp" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.647692 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.652495 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.653121 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.681044 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.686408 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/59bfeb19-1a15-4cbd-8970-5c18ff149992-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s9gvm\" (UID: \"59bfeb19-1a15-4cbd-8970-5c18ff149992\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.822270 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.843619 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2"] Oct 14 08:01:47 crc kubenswrapper[5002]: W1014 08:01:47.850030 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb6927e5_34dd_43d1_9dea_db00339df27e.slice/crio-90eda134e81a878d63d71cb03988727646d24e14160fee4f2509a3a4089e9386 WatchSource:0}: Error finding container 90eda134e81a878d63d71cb03988727646d24e14160fee4f2509a3a4089e9386: Status 404 returned error can't find the container with id 90eda134e81a878d63d71cb03988727646d24e14160fee4f2509a3a4089e9386 Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.885676 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8"] Oct 14 08:01:47 crc kubenswrapper[5002]: W1014 08:01:47.891512 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89f2e467_845f_45fb_b411_48c0cf2d3f0d.slice/crio-46d28ea043d73dbd37bcbe387dd41042970f7d1f5d5ff8e4348723a10b025df0 WatchSource:0}: Error finding container 46d28ea043d73dbd37bcbe387dd41042970f7d1f5d5ff8e4348723a10b025df0: Status 404 returned error can't find the container with id 46d28ea043d73dbd37bcbe387dd41042970f7d1f5d5ff8e4348723a10b025df0 Oct 14 08:01:47 crc kubenswrapper[5002]: I1014 08:01:47.971857 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-69797c4689-gjj84"] Oct 14 08:01:47 crc kubenswrapper[5002]: W1014 08:01:47.976753 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50e01b14_76e1_4d96_bc0e_38e89c3061ac.slice/crio-2219de8de968fe3087c4ecad754279f8502f1ced384440462bda308f37ca3d41 WatchSource:0}: Error finding container 2219de8de968fe3087c4ecad754279f8502f1ced384440462bda308f37ca3d41: Status 404 returned error can't find the container with id 2219de8de968fe3087c4ecad754279f8502f1ced384440462bda308f37ca3d41 Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:47.999427 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm"] Oct 14 08:01:48 crc kubenswrapper[5002]: W1014 08:01:48.005158 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59bfeb19_1a15_4cbd_8970_5c18ff149992.slice/crio-45d0a66e40ea3faf4a989513026c076905cfe69405565c56150b66d92d8cfc2d WatchSource:0}: Error finding container 45d0a66e40ea3faf4a989513026c076905cfe69405565c56150b66d92d8cfc2d: Status 404 returned error can't find the container with id 45d0a66e40ea3faf4a989513026c076905cfe69405565c56150b66d92d8cfc2d Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.480137 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69797c4689-gjj84" event={"ID":"50e01b14-76e1-4d96-bc0e-38e89c3061ac","Type":"ContainerStarted","Data":"c8ec54b52d6aedfc66e18efbd8bbcd5e446efc24218f5dfff04f2e1ae2e51c2d"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.480581 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-69797c4689-gjj84" event={"ID":"50e01b14-76e1-4d96-bc0e-38e89c3061ac","Type":"ContainerStarted","Data":"2219de8de968fe3087c4ecad754279f8502f1ced384440462bda308f37ca3d41"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.484119 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" event={"ID":"59bfeb19-1a15-4cbd-8970-5c18ff149992","Type":"ContainerStarted","Data":"45d0a66e40ea3faf4a989513026c076905cfe69405565c56150b66d92d8cfc2d"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.488119 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" event={"ID":"cb6927e5-34dd-43d1-9dea-db00339df27e","Type":"ContainerStarted","Data":"90eda134e81a878d63d71cb03988727646d24e14160fee4f2509a3a4089e9386"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.493981 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-82rj9" event={"ID":"0629a9d2-3340-4e00-9673-baaaf354a8ce","Type":"ContainerStarted","Data":"9995589a71cb2efa2cb786fe21452814e4f8c066a26a661701489687728f52d3"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.496973 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" event={"ID":"89f2e467-845f-45fb-b411-48c0cf2d3f0d","Type":"ContainerStarted","Data":"46d28ea043d73dbd37bcbe387dd41042970f7d1f5d5ff8e4348723a10b025df0"} Oct 14 08:01:48 crc kubenswrapper[5002]: I1014 08:01:48.515812 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-69797c4689-gjj84" podStartSLOduration=1.515784099 podStartE2EDuration="1.515784099s" podCreationTimestamp="2025-10-14 08:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:01:48.513876652 +0000 UTC m=+641.495116134" watchObservedRunningTime="2025-10-14 08:01:48.515784099 +0000 UTC m=+641.497023581" Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.525131 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" event={"ID":"59bfeb19-1a15-4cbd-8970-5c18ff149992","Type":"ContainerStarted","Data":"25b6f97ca5f1da6cc9ffb07f08cef26d4f87236524431d397d931abfb07d2473"} Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.526684 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" event={"ID":"cb6927e5-34dd-43d1-9dea-db00339df27e","Type":"ContainerStarted","Data":"f5274055c820a8afba574f532cc0bbf47587e12f5c8cd90a64b3c3d95e5c2724"} Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.526833 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.529425 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-82rj9" event={"ID":"0629a9d2-3340-4e00-9673-baaaf354a8ce","Type":"ContainerStarted","Data":"4f9faf4396acc1d58cb062129ba978272cb81fec3f408c0bbb8875d622631335"} Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.529676 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.531601 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" event={"ID":"89f2e467-845f-45fb-b411-48c0cf2d3f0d","Type":"ContainerStarted","Data":"6e99d5edccd02eabc9e429bb60555a344e4b6cb3aa706ca14b8841d5c4f8e65c"} Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.559217 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s9gvm" podStartSLOduration=2.843871121 podStartE2EDuration="5.559190835s" podCreationTimestamp="2025-10-14 08:01:46 +0000 UTC" firstStartedPulling="2025-10-14 08:01:48.00722174 +0000 UTC m=+640.988461202" lastFinishedPulling="2025-10-14 08:01:50.722541424 +0000 UTC m=+643.703780916" observedRunningTime="2025-10-14 08:01:51.550078235 +0000 UTC m=+644.531317757" watchObservedRunningTime="2025-10-14 08:01:51.559190835 +0000 UTC m=+644.540430327" Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.596542 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" podStartSLOduration=2.709996074 podStartE2EDuration="5.596519344s" podCreationTimestamp="2025-10-14 08:01:46 +0000 UTC" firstStartedPulling="2025-10-14 08:01:47.851769511 +0000 UTC m=+640.833008963" lastFinishedPulling="2025-10-14 08:01:50.738292751 +0000 UTC m=+643.719532233" observedRunningTime="2025-10-14 08:01:51.567672448 +0000 UTC m=+644.548911910" watchObservedRunningTime="2025-10-14 08:01:51.596519344 +0000 UTC m=+644.577758826" Oct 14 08:01:51 crc kubenswrapper[5002]: I1014 08:01:51.603407 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-82rj9" podStartSLOduration=2.603985389 podStartE2EDuration="5.603395417s" podCreationTimestamp="2025-10-14 08:01:46 +0000 UTC" firstStartedPulling="2025-10-14 08:01:47.729159938 +0000 UTC m=+640.710399400" lastFinishedPulling="2025-10-14 08:01:50.728569946 +0000 UTC m=+643.709809428" observedRunningTime="2025-10-14 08:01:51.586256655 +0000 UTC m=+644.567496117" watchObservedRunningTime="2025-10-14 08:01:51.603395417 +0000 UTC m=+644.584634909" Oct 14 08:01:54 crc kubenswrapper[5002]: I1014 08:01:54.559533 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" event={"ID":"89f2e467-845f-45fb-b411-48c0cf2d3f0d","Type":"ContainerStarted","Data":"55f2a4c6118eef77a216567b4becda10c9e73e3337cc312ea2c9c648f262e192"} Oct 14 08:01:54 crc kubenswrapper[5002]: I1014 08:01:54.594694 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dq7k8" podStartSLOduration=3.119658455 podStartE2EDuration="8.59466316s" podCreationTimestamp="2025-10-14 08:01:46 +0000 UTC" firstStartedPulling="2025-10-14 08:01:47.893997463 +0000 UTC m=+640.875236915" lastFinishedPulling="2025-10-14 08:01:53.369002128 +0000 UTC m=+646.350241620" observedRunningTime="2025-10-14 08:01:54.584038792 +0000 UTC m=+647.565278334" watchObservedRunningTime="2025-10-14 08:01:54.59466316 +0000 UTC m=+647.575902622" Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.456832 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.457284 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.465124 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.588495 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-69797c4689-gjj84" Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.664047 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 08:01:57 crc kubenswrapper[5002]: I1014 08:01:57.701561 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-82rj9" Oct 14 08:02:07 crc kubenswrapper[5002]: I1014 08:02:07.656814 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4tmr2" Oct 14 08:02:22 crc kubenswrapper[5002]: I1014 08:02:22.735500 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sn4pd" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerName="console" containerID="cri-o://1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8" gracePeriod=15 Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.197314 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sn4pd_6389e00b-ba44-426e-b07a-97e18189d9ef/console/0.log" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.197768 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294413 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294467 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lqn9\" (UniqueName: \"kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294526 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294552 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294588 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294615 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.294741 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config\") pod \"6389e00b-ba44-426e-b07a-97e18189d9ef\" (UID: \"6389e00b-ba44-426e-b07a-97e18189d9ef\") " Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.295689 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config" (OuterVolumeSpecName: "console-config") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.296003 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.296044 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca" (OuterVolumeSpecName: "service-ca") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.296154 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.301708 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.306148 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9" (OuterVolumeSpecName: "kube-api-access-4lqn9") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "kube-api-access-4lqn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.307207 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6389e00b-ba44-426e-b07a-97e18189d9ef" (UID: "6389e00b-ba44-426e-b07a-97e18189d9ef"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396089 5002 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396120 5002 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396131 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lqn9\" (UniqueName: \"kubernetes.io/projected/6389e00b-ba44-426e-b07a-97e18189d9ef-kube-api-access-4lqn9\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396141 5002 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396150 5002 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6389e00b-ba44-426e-b07a-97e18189d9ef-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396158 5002 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.396166 5002 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6389e00b-ba44-426e-b07a-97e18189d9ef-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769030 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sn4pd_6389e00b-ba44-426e-b07a-97e18189d9ef/console/0.log" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769082 5002 generic.go:334] "Generic (PLEG): container finished" podID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerID="1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8" exitCode=2 Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769115 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sn4pd" event={"ID":"6389e00b-ba44-426e-b07a-97e18189d9ef","Type":"ContainerDied","Data":"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8"} Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769142 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sn4pd" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769152 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sn4pd" event={"ID":"6389e00b-ba44-426e-b07a-97e18189d9ef","Type":"ContainerDied","Data":"bf1b657658355473476a75e25fb6ce6bd2541d5079dee274d837a81120005bd2"} Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.769181 5002 scope.go:117] "RemoveContainer" containerID="1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.787921 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.790200 5002 scope.go:117] "RemoveContainer" containerID="1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8" Oct 14 08:02:23 crc kubenswrapper[5002]: E1014 08:02:23.790724 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8\": container with ID starting with 1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8 not found: ID does not exist" containerID="1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.790816 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8"} err="failed to get container status \"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8\": rpc error: code = NotFound desc = could not find container \"1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8\": container with ID starting with 1ba892d1d8791f8a101f30403d03e5a94d473a576745d3199138f199b3ccf2b8 not found: ID does not exist" Oct 14 08:02:23 crc kubenswrapper[5002]: I1014 08:02:23.793102 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sn4pd"] Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.888893 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv"] Oct 14 08:02:24 crc kubenswrapper[5002]: E1014 08:02:24.889213 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerName="console" Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.889233 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerName="console" Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.889401 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" containerName="console" Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.891041 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.894965 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 08:02:24 crc kubenswrapper[5002]: I1014 08:02:24.918831 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv"] Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.018030 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtp6x\" (UniqueName: \"kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.018387 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.018661 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.120280 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtp6x\" (UniqueName: \"kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.120358 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.120451 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.121865 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.122323 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.162122 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtp6x\" (UniqueName: \"kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.223074 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.495072 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv"] Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.753950 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6389e00b-ba44-426e-b07a-97e18189d9ef" path="/var/lib/kubelet/pods/6389e00b-ba44-426e-b07a-97e18189d9ef/volumes" Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.790910 5002 generic.go:334] "Generic (PLEG): container finished" podID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerID="ba07ec3a55dfea9f7320ebeac856328992f00bd1917f5d2d25bcb72060a48591" exitCode=0 Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.790976 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" event={"ID":"9a92c18c-a9d6-470f-9236-feab1741aa53","Type":"ContainerDied","Data":"ba07ec3a55dfea9f7320ebeac856328992f00bd1917f5d2d25bcb72060a48591"} Oct 14 08:02:25 crc kubenswrapper[5002]: I1014 08:02:25.791016 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" event={"ID":"9a92c18c-a9d6-470f-9236-feab1741aa53","Type":"ContainerStarted","Data":"a72507b75d35c90aa9f828353e00f7545618e9d8ca1fba6a3a2b15a425735cdb"} Oct 14 08:02:27 crc kubenswrapper[5002]: I1014 08:02:27.816057 5002 generic.go:334] "Generic (PLEG): container finished" podID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerID="5a4fbba1e49c16dc16da580a82eed26ada858ec5f3190b5e6c98de7c3b717ddd" exitCode=0 Oct 14 08:02:27 crc kubenswrapper[5002]: I1014 08:02:27.816172 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" event={"ID":"9a92c18c-a9d6-470f-9236-feab1741aa53","Type":"ContainerDied","Data":"5a4fbba1e49c16dc16da580a82eed26ada858ec5f3190b5e6c98de7c3b717ddd"} Oct 14 08:02:28 crc kubenswrapper[5002]: I1014 08:02:28.826225 5002 generic.go:334] "Generic (PLEG): container finished" podID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerID="341d28fb27263f10026aff00fb097deb098777d67516c01a9c683ce2039ba3b4" exitCode=0 Oct 14 08:02:28 crc kubenswrapper[5002]: I1014 08:02:28.826294 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" event={"ID":"9a92c18c-a9d6-470f-9236-feab1741aa53","Type":"ContainerDied","Data":"341d28fb27263f10026aff00fb097deb098777d67516c01a9c683ce2039ba3b4"} Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.177673 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.304078 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtp6x\" (UniqueName: \"kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x\") pod \"9a92c18c-a9d6-470f-9236-feab1741aa53\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.304163 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util\") pod \"9a92c18c-a9d6-470f-9236-feab1741aa53\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.304228 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle\") pod \"9a92c18c-a9d6-470f-9236-feab1741aa53\" (UID: \"9a92c18c-a9d6-470f-9236-feab1741aa53\") " Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.305374 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle" (OuterVolumeSpecName: "bundle") pod "9a92c18c-a9d6-470f-9236-feab1741aa53" (UID: "9a92c18c-a9d6-470f-9236-feab1741aa53"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.312203 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x" (OuterVolumeSpecName: "kube-api-access-xtp6x") pod "9a92c18c-a9d6-470f-9236-feab1741aa53" (UID: "9a92c18c-a9d6-470f-9236-feab1741aa53"). InnerVolumeSpecName "kube-api-access-xtp6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.338527 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util" (OuterVolumeSpecName: "util") pod "9a92c18c-a9d6-470f-9236-feab1741aa53" (UID: "9a92c18c-a9d6-470f-9236-feab1741aa53"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.405940 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtp6x\" (UniqueName: \"kubernetes.io/projected/9a92c18c-a9d6-470f-9236-feab1741aa53-kube-api-access-xtp6x\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.406101 5002 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-util\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.406123 5002 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9a92c18c-a9d6-470f-9236-feab1741aa53-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.844864 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" event={"ID":"9a92c18c-a9d6-470f-9236-feab1741aa53","Type":"ContainerDied","Data":"a72507b75d35c90aa9f828353e00f7545618e9d8ca1fba6a3a2b15a425735cdb"} Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.845199 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72507b75d35c90aa9f828353e00f7545618e9d8ca1fba6a3a2b15a425735cdb" Oct 14 08:02:30 crc kubenswrapper[5002]: I1014 08:02:30.844938 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv" Oct 14 08:02:39 crc kubenswrapper[5002]: I1014 08:02:39.218269 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:02:39 crc kubenswrapper[5002]: I1014 08:02:39.218882 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.402581 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z"] Oct 14 08:02:40 crc kubenswrapper[5002]: E1014 08:02:40.402781 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="pull" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.402792 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="pull" Oct 14 08:02:40 crc kubenswrapper[5002]: E1014 08:02:40.402806 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="util" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.402812 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="util" Oct 14 08:02:40 crc kubenswrapper[5002]: E1014 08:02:40.402830 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="extract" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.402853 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="extract" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.402954 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a92c18c-a9d6-470f-9236-feab1741aa53" containerName="extract" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.403290 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.406474 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.406496 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.407091 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-48gwm" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.408245 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.408447 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.425308 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z"] Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.550215 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-webhook-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.550298 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk8tz\" (UniqueName: \"kubernetes.io/projected/45cb6486-ddee-4491-86d3-6de31fa2eeed-kube-api-access-jk8tz\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.550497 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-apiservice-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.651756 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk8tz\" (UniqueName: \"kubernetes.io/projected/45cb6486-ddee-4491-86d3-6de31fa2eeed-kube-api-access-jk8tz\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.652064 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-apiservice-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.652241 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-webhook-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.657474 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-webhook-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.657633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45cb6486-ddee-4491-86d3-6de31fa2eeed-apiservice-cert\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.670447 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk8tz\" (UniqueName: \"kubernetes.io/projected/45cb6486-ddee-4491-86d3-6de31fa2eeed-kube-api-access-jk8tz\") pod \"metallb-operator-controller-manager-7769bf444d-r6x5z\" (UID: \"45cb6486-ddee-4491-86d3-6de31fa2eeed\") " pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.718374 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.745429 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62"] Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.746123 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.750289 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.750383 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.750498 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-89fxp" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.821644 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62"] Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.854883 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t25b4\" (UniqueName: \"kubernetes.io/projected/513044aa-d565-47b2-8fb7-5df37788a795-kube-api-access-t25b4\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.854979 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-apiservice-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.855044 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-webhook-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.955848 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-webhook-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.956170 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t25b4\" (UniqueName: \"kubernetes.io/projected/513044aa-d565-47b2-8fb7-5df37788a795-kube-api-access-t25b4\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.956202 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-apiservice-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.962669 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-apiservice-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.966580 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/513044aa-d565-47b2-8fb7-5df37788a795-webhook-cert\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:40 crc kubenswrapper[5002]: I1014 08:02:40.980604 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t25b4\" (UniqueName: \"kubernetes.io/projected/513044aa-d565-47b2-8fb7-5df37788a795-kube-api-access-t25b4\") pod \"metallb-operator-webhook-server-74b78f565b-6dh62\" (UID: \"513044aa-d565-47b2-8fb7-5df37788a795\") " pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:41 crc kubenswrapper[5002]: I1014 08:02:41.054752 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z"] Oct 14 08:02:41 crc kubenswrapper[5002]: W1014 08:02:41.068487 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45cb6486_ddee_4491_86d3_6de31fa2eeed.slice/crio-15aa3dc5ed4451a8e1e0e369230d6bab401da7ea01aa9c5c906f59865d5ae32b WatchSource:0}: Error finding container 15aa3dc5ed4451a8e1e0e369230d6bab401da7ea01aa9c5c906f59865d5ae32b: Status 404 returned error can't find the container with id 15aa3dc5ed4451a8e1e0e369230d6bab401da7ea01aa9c5c906f59865d5ae32b Oct 14 08:02:41 crc kubenswrapper[5002]: I1014 08:02:41.084080 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:41 crc kubenswrapper[5002]: I1014 08:02:41.516875 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62"] Oct 14 08:02:41 crc kubenswrapper[5002]: W1014 08:02:41.523860 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod513044aa_d565_47b2_8fb7_5df37788a795.slice/crio-d277ac1b3a48a13db731519a47620b1cfc0e7f285ae0ea1ff5e510ece033331d WatchSource:0}: Error finding container d277ac1b3a48a13db731519a47620b1cfc0e7f285ae0ea1ff5e510ece033331d: Status 404 returned error can't find the container with id d277ac1b3a48a13db731519a47620b1cfc0e7f285ae0ea1ff5e510ece033331d Oct 14 08:02:41 crc kubenswrapper[5002]: I1014 08:02:41.921212 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" event={"ID":"45cb6486-ddee-4491-86d3-6de31fa2eeed","Type":"ContainerStarted","Data":"15aa3dc5ed4451a8e1e0e369230d6bab401da7ea01aa9c5c906f59865d5ae32b"} Oct 14 08:02:41 crc kubenswrapper[5002]: I1014 08:02:41.924011 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" event={"ID":"513044aa-d565-47b2-8fb7-5df37788a795","Type":"ContainerStarted","Data":"d277ac1b3a48a13db731519a47620b1cfc0e7f285ae0ea1ff5e510ece033331d"} Oct 14 08:02:45 crc kubenswrapper[5002]: I1014 08:02:45.960895 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" event={"ID":"513044aa-d565-47b2-8fb7-5df37788a795","Type":"ContainerStarted","Data":"68de5dd65ea6ddd78f71ad6b07354cc0aa2a95f315e24b9e20044a3aa280828c"} Oct 14 08:02:45 crc kubenswrapper[5002]: I1014 08:02:45.961337 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:02:45 crc kubenswrapper[5002]: I1014 08:02:45.962108 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" event={"ID":"45cb6486-ddee-4491-86d3-6de31fa2eeed","Type":"ContainerStarted","Data":"0d4c16b53d9aa8b2909a9ad1d93557445d4deb5a835ba03ec11570a01bfa166f"} Oct 14 08:02:45 crc kubenswrapper[5002]: I1014 08:02:45.962247 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:02:45 crc kubenswrapper[5002]: I1014 08:02:45.983265 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" podStartSLOduration=1.799553676 podStartE2EDuration="5.983248745s" podCreationTimestamp="2025-10-14 08:02:40 +0000 UTC" firstStartedPulling="2025-10-14 08:02:41.526871363 +0000 UTC m=+694.508110815" lastFinishedPulling="2025-10-14 08:02:45.710566432 +0000 UTC m=+698.691805884" observedRunningTime="2025-10-14 08:02:45.981146761 +0000 UTC m=+698.962386223" watchObservedRunningTime="2025-10-14 08:02:45.983248745 +0000 UTC m=+698.964488197" Oct 14 08:02:46 crc kubenswrapper[5002]: I1014 08:02:46.005789 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" podStartSLOduration=1.405974651 podStartE2EDuration="6.005772622s" podCreationTimestamp="2025-10-14 08:02:40 +0000 UTC" firstStartedPulling="2025-10-14 08:02:41.074023228 +0000 UTC m=+694.055262680" lastFinishedPulling="2025-10-14 08:02:45.673821199 +0000 UTC m=+698.655060651" observedRunningTime="2025-10-14 08:02:46.003808613 +0000 UTC m=+698.985048075" watchObservedRunningTime="2025-10-14 08:02:46.005772622 +0000 UTC m=+698.987012064" Oct 14 08:03:01 crc kubenswrapper[5002]: I1014 08:03:01.091305 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-74b78f565b-6dh62" Oct 14 08:03:09 crc kubenswrapper[5002]: I1014 08:03:09.218065 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:03:09 crc kubenswrapper[5002]: I1014 08:03:09.218677 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:03:20 crc kubenswrapper[5002]: I1014 08:03:20.724147 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7769bf444d-r6x5z" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.553398 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.554492 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.557713 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-t5nqw"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.560180 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.564384 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.564459 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.568610 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.569089 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-j8nrs" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.573627 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625608 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625655 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-sockets\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625678 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg5hw\" (UniqueName: \"kubernetes.io/projected/10d5d438-0656-4699-b386-a539ca6810d2-kube-api-access-tg5hw\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625717 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnzm9\" (UniqueName: \"kubernetes.io/projected/a31644e2-8ecc-4402-8776-2eb5815c7d55-kube-api-access-fnzm9\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625753 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625771 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625800 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-reloader\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625849 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-conf\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.625873 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-startup\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.648088 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-686b6"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.649196 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.651014 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.651095 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.651158 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qjfzk" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.651777 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.666276 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-2zwcb"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.667298 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.668782 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.684179 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2zwcb"] Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.727992 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.728144 5002 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.728765 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert podName:10d5d438-0656-4699-b386-a539ca6810d2 nodeName:}" failed. No retries permitted until 2025-10-14 08:03:22.228744705 +0000 UTC m=+735.209984157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert") pod "frr-k8s-webhook-server-64bf5d555-lf7b4" (UID: "10d5d438-0656-4699-b386-a539ca6810d2") : secret "frr-k8s-webhook-server-cert" not found Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.728705 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-sockets\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.728904 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg5hw\" (UniqueName: \"kubernetes.io/projected/10d5d438-0656-4699-b386-a539ca6810d2-kube-api-access-tg5hw\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.728966 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metrics-certs\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729072 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metallb-excludel2\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729104 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnzm9\" (UniqueName: \"kubernetes.io/projected/a31644e2-8ecc-4402-8776-2eb5815c7d55-kube-api-access-fnzm9\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729143 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hp7v\" (UniqueName: \"kubernetes.io/projected/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-kube-api-access-7hp7v\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729187 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-metrics-certs\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729238 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729255 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qglp6\" (UniqueName: \"kubernetes.io/projected/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-kube-api-access-qglp6\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729272 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729317 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729348 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-reloader\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729389 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-cert\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729418 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-conf\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.729425 5002 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729444 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-startup\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.729475 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs podName:a31644e2-8ecc-4402-8776-2eb5815c7d55 nodeName:}" failed. No retries permitted until 2025-10-14 08:03:22.229458154 +0000 UTC m=+735.210697606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs") pod "frr-k8s-t5nqw" (UID: "a31644e2-8ecc-4402-8776-2eb5815c7d55") : secret "frr-k8s-certs-secret" not found Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.729912 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-sockets\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.730012 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.730082 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-conf\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.730161 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a31644e2-8ecc-4402-8776-2eb5815c7d55-reloader\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.730572 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a31644e2-8ecc-4402-8776-2eb5815c7d55-frr-startup\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.747946 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnzm9\" (UniqueName: \"kubernetes.io/projected/a31644e2-8ecc-4402-8776-2eb5815c7d55-kube-api-access-fnzm9\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.754392 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg5hw\" (UniqueName: \"kubernetes.io/projected/10d5d438-0656-4699-b386-a539ca6810d2-kube-api-access-tg5hw\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.830820 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metallb-excludel2\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831199 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hp7v\" (UniqueName: \"kubernetes.io/projected/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-kube-api-access-7hp7v\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831239 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-metrics-certs\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831278 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831299 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qglp6\" (UniqueName: \"kubernetes.io/projected/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-kube-api-access-qglp6\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831366 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-cert\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.831500 5002 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 08:03:21 crc kubenswrapper[5002]: E1014 08:03:21.831565 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist podName:6d62cfe7-6906-4f1f-bc0b-595c98d4af86 nodeName:}" failed. No retries permitted until 2025-10-14 08:03:22.331550341 +0000 UTC m=+735.312789793 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist") pod "speaker-686b6" (UID: "6d62cfe7-6906-4f1f-bc0b-595c98d4af86") : secret "metallb-memberlist" not found Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.831988 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metrics-certs\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.832675 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metallb-excludel2\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.833972 5002 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.835722 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-metrics-certs\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.836325 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-metrics-certs\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.845371 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-cert\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.849575 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hp7v\" (UniqueName: \"kubernetes.io/projected/0a00fbf0-fcf8-42f8-a8da-cbf266655ee3-kube-api-access-7hp7v\") pod \"controller-68d546b9d8-2zwcb\" (UID: \"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3\") " pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.853324 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qglp6\" (UniqueName: \"kubernetes.io/projected/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-kube-api-access-qglp6\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:21 crc kubenswrapper[5002]: I1014 08:03:21.984721 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.211117 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2zwcb"] Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.238004 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.238071 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.241374 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10d5d438-0656-4699-b386-a539ca6810d2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-lf7b4\" (UID: \"10d5d438-0656-4699-b386-a539ca6810d2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.242419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a31644e2-8ecc-4402-8776-2eb5815c7d55-metrics-certs\") pod \"frr-k8s-t5nqw\" (UID: \"a31644e2-8ecc-4402-8776-2eb5815c7d55\") " pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.339150 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:22 crc kubenswrapper[5002]: E1014 08:03:22.339288 5002 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 08:03:22 crc kubenswrapper[5002]: E1014 08:03:22.339340 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist podName:6d62cfe7-6906-4f1f-bc0b-595c98d4af86 nodeName:}" failed. No retries permitted until 2025-10-14 08:03:23.339324113 +0000 UTC m=+736.320563565 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist") pod "speaker-686b6" (UID: "6d62cfe7-6906-4f1f-bc0b-595c98d4af86") : secret "metallb-memberlist" not found Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.483430 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.497895 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:22 crc kubenswrapper[5002]: I1014 08:03:22.691542 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4"] Oct 14 08:03:22 crc kubenswrapper[5002]: W1014 08:03:22.699186 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10d5d438_0656_4699_b386_a539ca6810d2.slice/crio-81f908d4513009237aae19ffc17338d2471764b6dfa6a07a869ecde38ac86fe9 WatchSource:0}: Error finding container 81f908d4513009237aae19ffc17338d2471764b6dfa6a07a869ecde38ac86fe9: Status 404 returned error can't find the container with id 81f908d4513009237aae19ffc17338d2471764b6dfa6a07a869ecde38ac86fe9 Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.205305 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2zwcb" event={"ID":"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3","Type":"ContainerStarted","Data":"d128a5e692cef431b7d56316668f1f2877b9b68d2fe61a55430628335f0115dc"} Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.205370 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2zwcb" event={"ID":"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3","Type":"ContainerStarted","Data":"fa22970861b767c21f9cc2850a3f93328d4f212ae6a6cd431f0e941af1bef7b9"} Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.205390 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2zwcb" event={"ID":"0a00fbf0-fcf8-42f8-a8da-cbf266655ee3","Type":"ContainerStarted","Data":"c72e779556ab4d1d052bb0846173cc5c1b2bb42674228f96bea2128295d81a8d"} Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.205415 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.206333 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" event={"ID":"10d5d438-0656-4699-b386-a539ca6810d2","Type":"ContainerStarted","Data":"81f908d4513009237aae19ffc17338d2471764b6dfa6a07a869ecde38ac86fe9"} Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.207137 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"9317596bf5894be0facc664786080d0c827b131639b9c9dd1df1ef2a74034590"} Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.225925 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-2zwcb" podStartSLOduration=2.225906997 podStartE2EDuration="2.225906997s" podCreationTimestamp="2025-10-14 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:03:23.224701775 +0000 UTC m=+736.205941297" watchObservedRunningTime="2025-10-14 08:03:23.225906997 +0000 UTC m=+736.207146449" Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.354820 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.364483 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6d62cfe7-6906-4f1f-bc0b-595c98d4af86-memberlist\") pod \"speaker-686b6\" (UID: \"6d62cfe7-6906-4f1f-bc0b-595c98d4af86\") " pod="metallb-system/speaker-686b6" Oct 14 08:03:23 crc kubenswrapper[5002]: I1014 08:03:23.464909 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-686b6" Oct 14 08:03:24 crc kubenswrapper[5002]: I1014 08:03:24.228266 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-686b6" event={"ID":"6d62cfe7-6906-4f1f-bc0b-595c98d4af86","Type":"ContainerStarted","Data":"430eaa78b619bd0e9176ba7d56f24f7f1b8ae433a60c61942fdb8c578e114dfa"} Oct 14 08:03:24 crc kubenswrapper[5002]: I1014 08:03:24.228561 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-686b6" event={"ID":"6d62cfe7-6906-4f1f-bc0b-595c98d4af86","Type":"ContainerStarted","Data":"fdc27782d0bf11a4f3bf4d62eaefb9d6150c9b629bdf1988ed543e461b280d6c"} Oct 14 08:03:24 crc kubenswrapper[5002]: I1014 08:03:24.228577 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-686b6" event={"ID":"6d62cfe7-6906-4f1f-bc0b-595c98d4af86","Type":"ContainerStarted","Data":"d78d0a4e1385c2e75d4ee04737f07c457d75b2d363d4aa6d6e725d155eccdc4f"} Oct 14 08:03:24 crc kubenswrapper[5002]: I1014 08:03:24.228821 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-686b6" Oct 14 08:03:24 crc kubenswrapper[5002]: I1014 08:03:24.247999 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-686b6" podStartSLOduration=3.247909083 podStartE2EDuration="3.247909083s" podCreationTimestamp="2025-10-14 08:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:03:24.246118796 +0000 UTC m=+737.227358248" watchObservedRunningTime="2025-10-14 08:03:24.247909083 +0000 UTC m=+737.229148535" Oct 14 08:03:30 crc kubenswrapper[5002]: I1014 08:03:30.273054 5002 generic.go:334] "Generic (PLEG): container finished" podID="a31644e2-8ecc-4402-8776-2eb5815c7d55" containerID="26bdf48390aed8c90171f0736b12de6778606949ccafd018e1561290439ff08d" exitCode=0 Oct 14 08:03:30 crc kubenswrapper[5002]: I1014 08:03:30.273149 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerDied","Data":"26bdf48390aed8c90171f0736b12de6778606949ccafd018e1561290439ff08d"} Oct 14 08:03:30 crc kubenswrapper[5002]: I1014 08:03:30.278538 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" event={"ID":"10d5d438-0656-4699-b386-a539ca6810d2","Type":"ContainerStarted","Data":"a8de3c2ed932bbbf4f4191db0ee88beb2bbac0d2bc93166e6f73f1db9e6da276"} Oct 14 08:03:30 crc kubenswrapper[5002]: I1014 08:03:30.278732 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:30 crc kubenswrapper[5002]: I1014 08:03:30.348920 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" podStartSLOduration=2.330792278 podStartE2EDuration="9.348885668s" podCreationTimestamp="2025-10-14 08:03:21 +0000 UTC" firstStartedPulling="2025-10-14 08:03:22.701290396 +0000 UTC m=+735.682529848" lastFinishedPulling="2025-10-14 08:03:29.719383736 +0000 UTC m=+742.700623238" observedRunningTime="2025-10-14 08:03:30.337928246 +0000 UTC m=+743.319167798" watchObservedRunningTime="2025-10-14 08:03:30.348885668 +0000 UTC m=+743.330125160" Oct 14 08:03:31 crc kubenswrapper[5002]: I1014 08:03:31.289268 5002 generic.go:334] "Generic (PLEG): container finished" podID="a31644e2-8ecc-4402-8776-2eb5815c7d55" containerID="2d4334b47feaa2d7779318fc1b33aeee3bdde52597355eb311ab1a57a2860a02" exitCode=0 Oct 14 08:03:31 crc kubenswrapper[5002]: I1014 08:03:31.289349 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerDied","Data":"2d4334b47feaa2d7779318fc1b33aeee3bdde52597355eb311ab1a57a2860a02"} Oct 14 08:03:32 crc kubenswrapper[5002]: I1014 08:03:32.298675 5002 generic.go:334] "Generic (PLEG): container finished" podID="a31644e2-8ecc-4402-8776-2eb5815c7d55" containerID="47d07199acb01633c3766ef9a7d171c96db8cab855932eff9dfeb72e6e6c51b9" exitCode=0 Oct 14 08:03:32 crc kubenswrapper[5002]: I1014 08:03:32.298808 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerDied","Data":"47d07199acb01633c3766ef9a7d171c96db8cab855932eff9dfeb72e6e6c51b9"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.313373 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"7457a788e33bce7178b5d281de6860a7d5356b347103c6289ae4a47ef3641645"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.313697 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"125d7d9efbf0686062e58c503b915652a98953c82877953ba9e140c7671e51cc"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.313711 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"5283c4e2cee2bfcef8f900f79f8bcfd88f3ef99fca3b323d121bd641e3d67e7d"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.313722 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"e039f34a638e66bfa77f7a25da5107128c51dbb249fbb6d996a280ade99c9831"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.313733 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"7197f32431fd9516c88a2da3fc448c1e5a61c3577037ed1dd9e0f66e2f9ed497"} Oct 14 08:03:33 crc kubenswrapper[5002]: I1014 08:03:33.470699 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-686b6" Oct 14 08:03:34 crc kubenswrapper[5002]: I1014 08:03:34.327409 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-t5nqw" event={"ID":"a31644e2-8ecc-4402-8776-2eb5815c7d55","Type":"ContainerStarted","Data":"dcc48bf7df55b1f843ccd3677566cbbdf6b14998f79e21e7af8b244e0b5f1352"} Oct 14 08:03:34 crc kubenswrapper[5002]: I1014 08:03:34.328224 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:34 crc kubenswrapper[5002]: I1014 08:03:34.354025 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-t5nqw" podStartSLOduration=6.303408694 podStartE2EDuration="13.354004449s" podCreationTimestamp="2025-10-14 08:03:21 +0000 UTC" firstStartedPulling="2025-10-14 08:03:22.623553237 +0000 UTC m=+735.604792689" lastFinishedPulling="2025-10-14 08:03:29.674148982 +0000 UTC m=+742.655388444" observedRunningTime="2025-10-14 08:03:34.351898444 +0000 UTC m=+747.333137976" watchObservedRunningTime="2025-10-14 08:03:34.354004449 +0000 UTC m=+747.335243901" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.257735 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.258282 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerName="controller-manager" containerID="cri-o://8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9" gracePeriod=30 Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.351346 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.351559 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" podUID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" containerName="route-controller-manager" containerID="cri-o://6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367" gracePeriod=30 Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.696071 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.728261 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krmgs\" (UniqueName: \"kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs\") pod \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.732979 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles\") pod \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.733056 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert\") pod \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.733132 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config\") pod \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.733187 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca\") pod \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\" (UID: \"47b9ba60-7de6-4bd5-bb0f-edcf124193ad\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.734119 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config" (OuterVolumeSpecName: "config") pod "47b9ba60-7de6-4bd5-bb0f-edcf124193ad" (UID: "47b9ba60-7de6-4bd5-bb0f-edcf124193ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.734940 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "47b9ba60-7de6-4bd5-bb0f-edcf124193ad" (UID: "47b9ba60-7de6-4bd5-bb0f-edcf124193ad"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.736388 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca" (OuterVolumeSpecName: "client-ca") pod "47b9ba60-7de6-4bd5-bb0f-edcf124193ad" (UID: "47b9ba60-7de6-4bd5-bb0f-edcf124193ad"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.743263 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "47b9ba60-7de6-4bd5-bb0f-edcf124193ad" (UID: "47b9ba60-7de6-4bd5-bb0f-edcf124193ad"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.744282 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.744334 5002 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.744350 5002 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.751045 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs" (OuterVolumeSpecName: "kube-api-access-krmgs") pod "47b9ba60-7de6-4bd5-bb0f-edcf124193ad" (UID: "47b9ba60-7de6-4bd5-bb0f-edcf124193ad"). InnerVolumeSpecName "kube-api-access-krmgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.785448 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.845320 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config\") pod \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.845460 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkbmz\" (UniqueName: \"kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz\") pod \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.845541 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca\") pod \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.845596 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert\") pod \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\" (UID: \"17bc792f-98b0-46b1-b95c-00f3ee5e40a5\") " Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.846008 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krmgs\" (UniqueName: \"kubernetes.io/projected/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-kube-api-access-krmgs\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.846025 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47b9ba60-7de6-4bd5-bb0f-edcf124193ad-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.846057 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config" (OuterVolumeSpecName: "config") pod "17bc792f-98b0-46b1-b95c-00f3ee5e40a5" (UID: "17bc792f-98b0-46b1-b95c-00f3ee5e40a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.846758 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca" (OuterVolumeSpecName: "client-ca") pod "17bc792f-98b0-46b1-b95c-00f3ee5e40a5" (UID: "17bc792f-98b0-46b1-b95c-00f3ee5e40a5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.854981 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz" (OuterVolumeSpecName: "kube-api-access-bkbmz") pod "17bc792f-98b0-46b1-b95c-00f3ee5e40a5" (UID: "17bc792f-98b0-46b1-b95c-00f3ee5e40a5"). InnerVolumeSpecName "kube-api-access-bkbmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.859130 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "17bc792f-98b0-46b1-b95c-00f3ee5e40a5" (UID: "17bc792f-98b0-46b1-b95c-00f3ee5e40a5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.947533 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkbmz\" (UniqueName: \"kubernetes.io/projected/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-kube-api-access-bkbmz\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.947570 5002 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.947579 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:35 crc kubenswrapper[5002]: I1014 08:03:35.947588 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17bc792f-98b0-46b1-b95c-00f3ee5e40a5-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.345005 5002 generic.go:334] "Generic (PLEG): container finished" podID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" containerID="6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367" exitCode=0 Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.345084 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" event={"ID":"17bc792f-98b0-46b1-b95c-00f3ee5e40a5","Type":"ContainerDied","Data":"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367"} Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.345126 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.345464 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc" event={"ID":"17bc792f-98b0-46b1-b95c-00f3ee5e40a5","Type":"ContainerDied","Data":"1167936feb200e580244d138c63988345a1cb35d7555a19c1576867cc7a629af"} Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.345601 5002 scope.go:117] "RemoveContainer" containerID="6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.348214 5002 generic.go:334] "Generic (PLEG): container finished" podID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerID="8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9" exitCode=0 Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.348252 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.348261 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" event={"ID":"47b9ba60-7de6-4bd5-bb0f-edcf124193ad","Type":"ContainerDied","Data":"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9"} Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.348293 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hbt2w" event={"ID":"47b9ba60-7de6-4bd5-bb0f-edcf124193ad","Type":"ContainerDied","Data":"ab58f7a574edc35ca3602108e35435246e577800971724bcbf0c6baedf9db504"} Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.378890 5002 scope.go:117] "RemoveContainer" containerID="6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.382897 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 08:03:36 crc kubenswrapper[5002]: E1014 08:03:36.385297 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367\": container with ID starting with 6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367 not found: ID does not exist" containerID="6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.385391 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367"} err="failed to get container status \"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367\": rpc error: code = NotFound desc = could not find container \"6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367\": container with ID starting with 6e6c7196000fbdbff0958b6e961f815fa409f02be05fa715b5b80a4f24088367 not found: ID does not exist" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.385461 5002 scope.go:117] "RemoveContainer" containerID="8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.386757 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hbt2w"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.423405 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:36 crc kubenswrapper[5002]: E1014 08:03:36.423672 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerName="controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.423693 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerName="controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: E1014 08:03:36.423725 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" containerName="route-controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.423736 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" containerName="route-controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.423878 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" containerName="controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.423893 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" containerName="route-controller-manager" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.424318 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.427472 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.429185 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.431403 5002 scope.go:117] "RemoveContainer" containerID="8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9" Oct 14 08:03:36 crc kubenswrapper[5002]: E1014 08:03:36.432141 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9\": container with ID starting with 8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9 not found: ID does not exist" containerID="8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.432175 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9"} err="failed to get container status \"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9\": rpc error: code = NotFound desc = could not find container \"8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9\": container with ID starting with 8a9ef3039c158fc7ea688e433b91bd64b0943c655a2f43dbb04244669fe535f9 not found: ID does not exist" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.436657 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.440644 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mh8hc"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.483146 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.554578 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zsl\" (UniqueName: \"kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl\") pod \"openstack-operator-index-lbcjz\" (UID: \"b7fd79ac-4dd5-480b-ac37-fb77d6148345\") " pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.590223 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.591572 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.594459 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f88db758f-6bclk"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.595367 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.597516 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.597918 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.598140 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.600678 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.600743 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.600877 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.601051 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.601139 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.601553 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.601673 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.602041 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.602365 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.612126 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f88db758f-6bclk"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.613699 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.617923 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.655692 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-client-ca\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.655802 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.655955 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656062 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zsl\" (UniqueName: \"kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl\") pod \"openstack-operator-index-lbcjz\" (UID: \"b7fd79ac-4dd5-480b-ac37-fb77d6148345\") " pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656110 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x896v\" (UniqueName: \"kubernetes.io/projected/4d1cabbf-7f71-45ba-88a5-557ae7484742-kube-api-access-x896v\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656141 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1cabbf-7f71-45ba-88a5-557ae7484742-serving-cert\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656183 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-config\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656296 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-proxy-ca-bundles\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656351 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.656420 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2g9w\" (UniqueName: \"kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.684686 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zsl\" (UniqueName: \"kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl\") pod \"openstack-operator-index-lbcjz\" (UID: \"b7fd79ac-4dd5-480b-ac37-fb77d6148345\") " pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.742316 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.757882 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.757948 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2g9w\" (UniqueName: \"kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758003 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-client-ca\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758029 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758059 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758085 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x896v\" (UniqueName: \"kubernetes.io/projected/4d1cabbf-7f71-45ba-88a5-557ae7484742-kube-api-access-x896v\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758106 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1cabbf-7f71-45ba-88a5-557ae7484742-serving-cert\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758134 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-config\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.758180 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-proxy-ca-bundles\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.759797 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-client-ca\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.759906 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-proxy-ca-bundles\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.760553 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.761009 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.761424 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d1cabbf-7f71-45ba-88a5-557ae7484742-config\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.764089 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d1cabbf-7f71-45ba-88a5-557ae7484742-serving-cert\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.765474 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.780445 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2g9w\" (UniqueName: \"kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w\") pod \"route-controller-manager-5d49d48b8b-pqkh9\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.783248 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x896v\" (UniqueName: \"kubernetes.io/projected/4d1cabbf-7f71-45ba-88a5-557ae7484742-kube-api-access-x896v\") pod \"controller-manager-7f88db758f-6bclk\" (UID: \"4d1cabbf-7f71-45ba-88a5-557ae7484742\") " pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.971291 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:36 crc kubenswrapper[5002]: I1014 08:03:36.977690 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:36 crc kubenswrapper[5002]: W1014 08:03:36.988796 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7fd79ac_4dd5_480b_ac37_fb77d6148345.slice/crio-e14d1e374c48ae1d35a4e8bad211e0dd08c18a39a9b2fad5bd23789539ae5459 WatchSource:0}: Error finding container e14d1e374c48ae1d35a4e8bad211e0dd08c18a39a9b2fad5bd23789539ae5459: Status 404 returned error can't find the container with id e14d1e374c48ae1d35a4e8bad211e0dd08c18a39a9b2fad5bd23789539ae5459 Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.021609 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.040804 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.360259 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lbcjz" event={"ID":"b7fd79ac-4dd5-480b-ac37-fb77d6148345","Type":"ContainerStarted","Data":"e14d1e374c48ae1d35a4e8bad211e0dd08c18a39a9b2fad5bd23789539ae5459"} Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.451124 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:37 crc kubenswrapper[5002]: W1014 08:03:37.456922 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf0ecf4b_c96c_4ee6_ba15_6f5a0934d742.slice/crio-7281b17cdcc071ea89947738c9bc5c163a7bc9009512ade3902e667a8c3ac1d1 WatchSource:0}: Error finding container 7281b17cdcc071ea89947738c9bc5c163a7bc9009512ade3902e667a8c3ac1d1: Status 404 returned error can't find the container with id 7281b17cdcc071ea89947738c9bc5c163a7bc9009512ade3902e667a8c3ac1d1 Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.498879 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.509664 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f88db758f-6bclk"] Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.556464 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.732179 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17bc792f-98b0-46b1-b95c-00f3ee5e40a5" path="/var/lib/kubelet/pods/17bc792f-98b0-46b1-b95c-00f3ee5e40a5/volumes" Oct 14 08:03:37 crc kubenswrapper[5002]: I1014 08:03:37.733044 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47b9ba60-7de6-4bd5-bb0f-edcf124193ad" path="/var/lib/kubelet/pods/47b9ba60-7de6-4bd5-bb0f-edcf124193ad/volumes" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.366957 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" event={"ID":"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742","Type":"ContainerStarted","Data":"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a"} Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.367263 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerName="route-controller-manager" containerID="cri-o://88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a" gracePeriod=30 Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.367267 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" event={"ID":"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742","Type":"ContainerStarted","Data":"7281b17cdcc071ea89947738c9bc5c163a7bc9009512ade3902e667a8c3ac1d1"} Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.370001 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.371885 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" event={"ID":"4d1cabbf-7f71-45ba-88a5-557ae7484742","Type":"ContainerStarted","Data":"7e441e477838fae0e675e1c059d89b4656f954260557e79a712cbed210e0034e"} Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.371946 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" event={"ID":"4d1cabbf-7f71-45ba-88a5-557ae7484742","Type":"ContainerStarted","Data":"2c01fa60f3d0f1a3e276b0a8fff5987353b782352c6f19f36697646c96734994"} Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.374512 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.376741 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lbcjz" event={"ID":"b7fd79ac-4dd5-480b-ac37-fb77d6148345","Type":"ContainerStarted","Data":"5cd8da4232c89c128325acc04893bd67c5f8f7029e8abb12a7615b266d4ca9fc"} Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.378037 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.383480 5002 patch_prober.go:28] interesting pod/route-controller-manager-5d49d48b8b-pqkh9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.52:8443/healthz\": EOF" start-of-body= Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.383548 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.52:8443/healthz\": EOF" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.396823 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" podStartSLOduration=3.396801113 podStartE2EDuration="3.396801113s" podCreationTimestamp="2025-10-14 08:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:03:38.385460262 +0000 UTC m=+751.366699764" watchObservedRunningTime="2025-10-14 08:03:38.396801113 +0000 UTC m=+751.378040585" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.428725 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lbcjz" podStartSLOduration=1.299917035 podStartE2EDuration="2.428703883s" podCreationTimestamp="2025-10-14 08:03:36 +0000 UTC" firstStartedPulling="2025-10-14 08:03:36.994342522 +0000 UTC m=+749.975581974" lastFinishedPulling="2025-10-14 08:03:38.12312937 +0000 UTC m=+751.104368822" observedRunningTime="2025-10-14 08:03:38.426565775 +0000 UTC m=+751.407805237" watchObservedRunningTime="2025-10-14 08:03:38.428703883 +0000 UTC m=+751.409943345" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.447070 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f88db758f-6bclk" podStartSLOduration=3.447054121 podStartE2EDuration="3.447054121s" podCreationTimestamp="2025-10-14 08:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:03:38.442290923 +0000 UTC m=+751.423530375" watchObservedRunningTime="2025-10-14 08:03:38.447054121 +0000 UTC m=+751.428293573" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.764884 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.805991 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf"] Oct 14 08:03:38 crc kubenswrapper[5002]: E1014 08:03:38.806258 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerName="route-controller-manager" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.806277 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerName="route-controller-manager" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.806433 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerName="route-controller-manager" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.806949 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.814137 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf"] Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.890122 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca\") pod \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.890180 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2g9w\" (UniqueName: \"kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w\") pod \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.890225 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert\") pod \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.890284 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config\") pod \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\" (UID: \"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742\") " Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.890685 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-client-ca\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.891874 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config" (OuterVolumeSpecName: "config") pod "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" (UID: "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.892008 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15773fbb-19c0-4e25-88fe-221f6a79e38d-serving-cert\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.891907 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca" (OuterVolumeSpecName: "client-ca") pod "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" (UID: "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.892108 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-config\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.892160 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws2kv\" (UniqueName: \"kubernetes.io/projected/15773fbb-19c0-4e25-88fe-221f6a79e38d-kube-api-access-ws2kv\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.892284 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.892299 5002 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.895645 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w" (OuterVolumeSpecName: "kube-api-access-m2g9w") pod "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" (UID: "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742"). InnerVolumeSpecName "kube-api-access-m2g9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.900158 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" (UID: "bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993243 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-config\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993384 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws2kv\" (UniqueName: \"kubernetes.io/projected/15773fbb-19c0-4e25-88fe-221f6a79e38d-kube-api-access-ws2kv\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993526 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-client-ca\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993593 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15773fbb-19c0-4e25-88fe-221f6a79e38d-serving-cert\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993681 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2g9w\" (UniqueName: \"kubernetes.io/projected/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-kube-api-access-m2g9w\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.993704 5002 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.996181 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-client-ca\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.997166 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15773fbb-19c0-4e25-88fe-221f6a79e38d-config\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:38 crc kubenswrapper[5002]: I1014 08:03:38.999494 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15773fbb-19c0-4e25-88fe-221f6a79e38d-serving-cert\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.023114 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws2kv\" (UniqueName: \"kubernetes.io/projected/15773fbb-19c0-4e25-88fe-221f6a79e38d-kube-api-access-ws2kv\") pod \"route-controller-manager-6878795c87-9wwnf\" (UID: \"15773fbb-19c0-4e25-88fe-221f6a79e38d\") " pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.129271 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.218580 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.218633 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.218692 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.219357 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.219416 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f" gracePeriod=600 Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.384743 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f" exitCode=0 Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.385367 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f"} Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.385406 5002 scope.go:117] "RemoveContainer" containerID="c50c7f093728349969b14fe849e643da393a744ba5f6757ca4639925880513e3" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.389333 5002 generic.go:334] "Generic (PLEG): container finished" podID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" containerID="88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a" exitCode=0 Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.389441 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" event={"ID":"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742","Type":"ContainerDied","Data":"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a"} Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.389492 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" event={"ID":"bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742","Type":"ContainerDied","Data":"7281b17cdcc071ea89947738c9bc5c163a7bc9009512ade3902e667a8c3ac1d1"} Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.389451 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.424191 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.428801 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d49d48b8b-pqkh9"] Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.540488 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf"] Oct 14 08:03:39 crc kubenswrapper[5002]: W1014 08:03:39.555068 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15773fbb_19c0_4e25_88fe_221f6a79e38d.slice/crio-8693edfeedb8c0bbf18a0cad4645f0974d6037d123f5cdd0a2f295e5b232c43b WatchSource:0}: Error finding container 8693edfeedb8c0bbf18a0cad4645f0974d6037d123f5cdd0a2f295e5b232c43b: Status 404 returned error can't find the container with id 8693edfeedb8c0bbf18a0cad4645f0974d6037d123f5cdd0a2f295e5b232c43b Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.666099 5002 scope.go:117] "RemoveContainer" containerID="88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.708668 5002 scope.go:117] "RemoveContainer" containerID="88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a" Oct 14 08:03:39 crc kubenswrapper[5002]: E1014 08:03:39.709202 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a\": container with ID starting with 88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a not found: ID does not exist" containerID="88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.709241 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a"} err="failed to get container status \"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a\": rpc error: code = NotFound desc = could not find container \"88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a\": container with ID starting with 88ac208b6586aafd5bf53db9052115dc0719a40b0ea1ba47cc7e9c2151937c7a not found: ID does not exist" Oct 14 08:03:39 crc kubenswrapper[5002]: I1014 08:03:39.732567 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742" path="/var/lib/kubelet/pods/bf0ecf4b-c96c-4ee6-ba15-6f5a0934d742/volumes" Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.399793 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" event={"ID":"15773fbb-19c0-4e25-88fe-221f6a79e38d","Type":"ContainerStarted","Data":"61f8f659886d8a23afc5415e1aace516dee4967d50e7cafd7299e18159188193"} Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.400148 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.400164 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" event={"ID":"15773fbb-19c0-4e25-88fe-221f6a79e38d","Type":"ContainerStarted","Data":"8693edfeedb8c0bbf18a0cad4645f0974d6037d123f5cdd0a2f295e5b232c43b"} Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.406151 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a"} Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.418112 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" Oct 14 08:03:40 crc kubenswrapper[5002]: I1014 08:03:40.431738 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6878795c87-9wwnf" podStartSLOduration=3.431720195 podStartE2EDuration="3.431720195s" podCreationTimestamp="2025-10-14 08:03:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:03:40.428174951 +0000 UTC m=+753.409414443" watchObservedRunningTime="2025-10-14 08:03:40.431720195 +0000 UTC m=+753.412959667" Oct 14 08:03:41 crc kubenswrapper[5002]: I1014 08:03:41.989804 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-2zwcb" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.195582 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.196164 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-lbcjz" podUID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" containerName="registry-server" containerID="cri-o://5cd8da4232c89c128325acc04893bd67c5f8f7029e8abb12a7615b266d4ca9fc" gracePeriod=2 Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.423943 5002 generic.go:334] "Generic (PLEG): container finished" podID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" containerID="5cd8da4232c89c128325acc04893bd67c5f8f7029e8abb12a7615b266d4ca9fc" exitCode=0 Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.424002 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lbcjz" event={"ID":"b7fd79ac-4dd5-480b-ac37-fb77d6148345","Type":"ContainerDied","Data":"5cd8da4232c89c128325acc04893bd67c5f8f7029e8abb12a7615b266d4ca9fc"} Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.489509 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-lf7b4" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.501222 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-t5nqw" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.601486 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pgk79"] Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.602647 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.609504 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5df8j" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.610689 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pgk79"] Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.619884 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.646335 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbr7d\" (UniqueName: \"kubernetes.io/projected/4f65b042-e91d-4c89-9aa4-52abc662f94f-kube-api-access-dbr7d\") pod \"openstack-operator-index-pgk79\" (UID: \"4f65b042-e91d-4c89-9aa4-52abc662f94f\") " pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.747395 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zsl\" (UniqueName: \"kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl\") pod \"b7fd79ac-4dd5-480b-ac37-fb77d6148345\" (UID: \"b7fd79ac-4dd5-480b-ac37-fb77d6148345\") " Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.747776 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbr7d\" (UniqueName: \"kubernetes.io/projected/4f65b042-e91d-4c89-9aa4-52abc662f94f-kube-api-access-dbr7d\") pod \"openstack-operator-index-pgk79\" (UID: \"4f65b042-e91d-4c89-9aa4-52abc662f94f\") " pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.759591 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl" (OuterVolumeSpecName: "kube-api-access-x7zsl") pod "b7fd79ac-4dd5-480b-ac37-fb77d6148345" (UID: "b7fd79ac-4dd5-480b-ac37-fb77d6148345"). InnerVolumeSpecName "kube-api-access-x7zsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.764330 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbr7d\" (UniqueName: \"kubernetes.io/projected/4f65b042-e91d-4c89-9aa4-52abc662f94f-kube-api-access-dbr7d\") pod \"openstack-operator-index-pgk79\" (UID: \"4f65b042-e91d-4c89-9aa4-52abc662f94f\") " pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.849343 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zsl\" (UniqueName: \"kubernetes.io/projected/b7fd79ac-4dd5-480b-ac37-fb77d6148345-kube-api-access-x7zsl\") on node \"crc\" DevicePath \"\"" Oct 14 08:03:42 crc kubenswrapper[5002]: I1014 08:03:42.943831 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.038127 5002 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.359205 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pgk79"] Oct 14 08:03:43 crc kubenswrapper[5002]: W1014 08:03:43.366629 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f65b042_e91d_4c89_9aa4_52abc662f94f.slice/crio-2f75d689d458bc284324ab844dc1b0015dfa8c08c8e36dc6b4d03bfabdee3503 WatchSource:0}: Error finding container 2f75d689d458bc284324ab844dc1b0015dfa8c08c8e36dc6b4d03bfabdee3503: Status 404 returned error can't find the container with id 2f75d689d458bc284324ab844dc1b0015dfa8c08c8e36dc6b4d03bfabdee3503 Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.440126 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pgk79" event={"ID":"4f65b042-e91d-4c89-9aa4-52abc662f94f","Type":"ContainerStarted","Data":"2f75d689d458bc284324ab844dc1b0015dfa8c08c8e36dc6b4d03bfabdee3503"} Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.442010 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lbcjz" event={"ID":"b7fd79ac-4dd5-480b-ac37-fb77d6148345","Type":"ContainerDied","Data":"e14d1e374c48ae1d35a4e8bad211e0dd08c18a39a9b2fad5bd23789539ae5459"} Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.442065 5002 scope.go:117] "RemoveContainer" containerID="5cd8da4232c89c128325acc04893bd67c5f8f7029e8abb12a7615b266d4ca9fc" Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.442071 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lbcjz" Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.466782 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.471530 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-lbcjz"] Oct 14 08:03:43 crc kubenswrapper[5002]: I1014 08:03:43.728088 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" path="/var/lib/kubelet/pods/b7fd79ac-4dd5-480b-ac37-fb77d6148345/volumes" Oct 14 08:03:44 crc kubenswrapper[5002]: I1014 08:03:44.451272 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pgk79" event={"ID":"4f65b042-e91d-4c89-9aa4-52abc662f94f","Type":"ContainerStarted","Data":"208ada6e6f1d441c3a6ee5b0fff13c75b0687f8e70e45502af9af1f8b763e5e7"} Oct 14 08:03:44 crc kubenswrapper[5002]: I1014 08:03:44.469514 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pgk79" podStartSLOduration=1.8931888799999999 podStartE2EDuration="2.469486895s" podCreationTimestamp="2025-10-14 08:03:42 +0000 UTC" firstStartedPulling="2025-10-14 08:03:43.368884417 +0000 UTC m=+756.350123869" lastFinishedPulling="2025-10-14 08:03:43.945182432 +0000 UTC m=+756.926421884" observedRunningTime="2025-10-14 08:03:44.46776256 +0000 UTC m=+757.449002012" watchObservedRunningTime="2025-10-14 08:03:44.469486895 +0000 UTC m=+757.450726377" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.425561 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:03:50 crc kubenswrapper[5002]: E1014 08:03:50.427537 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" containerName="registry-server" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.427642 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" containerName="registry-server" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.427934 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7fd79ac-4dd5-480b-ac37-fb77d6148345" containerName="registry-server" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.428971 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.451032 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.465743 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.465914 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v85vg\" (UniqueName: \"kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.466083 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.567786 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.568131 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v85vg\" (UniqueName: \"kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.568321 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.568954 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.569183 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.609473 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v85vg\" (UniqueName: \"kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg\") pod \"community-operators-rscf5\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:50 crc kubenswrapper[5002]: I1014 08:03:50.769304 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:03:51 crc kubenswrapper[5002]: I1014 08:03:51.260733 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:03:51 crc kubenswrapper[5002]: W1014 08:03:51.273693 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode723db1a_1996_415e_a42e_e497689c4194.slice/crio-fdd9a790ab4c211c99f4c7bb0b79645ad2cee3bbbc2551da91e655d05b6d936f WatchSource:0}: Error finding container fdd9a790ab4c211c99f4c7bb0b79645ad2cee3bbbc2551da91e655d05b6d936f: Status 404 returned error can't find the container with id fdd9a790ab4c211c99f4c7bb0b79645ad2cee3bbbc2551da91e655d05b6d936f Oct 14 08:03:51 crc kubenswrapper[5002]: I1014 08:03:51.507882 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerStarted","Data":"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067"} Oct 14 08:03:51 crc kubenswrapper[5002]: I1014 08:03:51.507951 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerStarted","Data":"fdd9a790ab4c211c99f4c7bb0b79645ad2cee3bbbc2551da91e655d05b6d936f"} Oct 14 08:03:52 crc kubenswrapper[5002]: I1014 08:03:52.520918 5002 generic.go:334] "Generic (PLEG): container finished" podID="e723db1a-1996-415e-a42e-e497689c4194" containerID="9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067" exitCode=0 Oct 14 08:03:52 crc kubenswrapper[5002]: I1014 08:03:52.520996 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerDied","Data":"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067"} Oct 14 08:03:52 crc kubenswrapper[5002]: I1014 08:03:52.944909 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:52 crc kubenswrapper[5002]: I1014 08:03:52.944966 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:52 crc kubenswrapper[5002]: I1014 08:03:52.974084 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:53 crc kubenswrapper[5002]: I1014 08:03:53.532539 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerStarted","Data":"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2"} Oct 14 08:03:53 crc kubenswrapper[5002]: I1014 08:03:53.563396 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-pgk79" Oct 14 08:03:54 crc kubenswrapper[5002]: I1014 08:03:54.542207 5002 generic.go:334] "Generic (PLEG): container finished" podID="e723db1a-1996-415e-a42e-e497689c4194" containerID="90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2" exitCode=0 Oct 14 08:03:54 crc kubenswrapper[5002]: I1014 08:03:54.542311 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerDied","Data":"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2"} Oct 14 08:03:55 crc kubenswrapper[5002]: I1014 08:03:55.549868 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerStarted","Data":"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af"} Oct 14 08:03:55 crc kubenswrapper[5002]: I1014 08:03:55.576162 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rscf5" podStartSLOduration=3.097443885 podStartE2EDuration="5.576143087s" podCreationTimestamp="2025-10-14 08:03:50 +0000 UTC" firstStartedPulling="2025-10-14 08:03:52.525115775 +0000 UTC m=+765.506355267" lastFinishedPulling="2025-10-14 08:03:55.003814977 +0000 UTC m=+767.985054469" observedRunningTime="2025-10-14 08:03:55.574976186 +0000 UTC m=+768.556215648" watchObservedRunningTime="2025-10-14 08:03:55.576143087 +0000 UTC m=+768.557382539" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.681346 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s"] Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.683729 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.686561 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-7tptm" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.703869 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s"] Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.796759 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.797189 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzqfp\" (UniqueName: \"kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.806094 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.908145 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.908326 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzqfp\" (UniqueName: \"kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.908401 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.909004 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.909429 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:03:59 crc kubenswrapper[5002]: I1014 08:03:59.931780 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzqfp\" (UniqueName: \"kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.005812 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.488073 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s"] Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.585264 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" event={"ID":"60f12264-cedf-4b85-8aa3-2707d34e6a00","Type":"ContainerStarted","Data":"2adc7957c1a7ca5520235d689d1f16511fa7ec3a12a4e0f628042814d889fef8"} Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.770041 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.770130 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:00 crc kubenswrapper[5002]: I1014 08:04:00.835492 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:01 crc kubenswrapper[5002]: I1014 08:04:01.596784 5002 generic.go:334] "Generic (PLEG): container finished" podID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerID="a65c4c5302aecb8f9633d27d74bc35ff2e6822ef57bdbc696971def8f62fd0ee" exitCode=0 Oct 14 08:04:01 crc kubenswrapper[5002]: I1014 08:04:01.596914 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" event={"ID":"60f12264-cedf-4b85-8aa3-2707d34e6a00","Type":"ContainerDied","Data":"a65c4c5302aecb8f9633d27d74bc35ff2e6822ef57bdbc696971def8f62fd0ee"} Oct 14 08:04:01 crc kubenswrapper[5002]: I1014 08:04:01.675575 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:03 crc kubenswrapper[5002]: I1014 08:04:03.394445 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:04:03 crc kubenswrapper[5002]: I1014 08:04:03.613418 5002 generic.go:334] "Generic (PLEG): container finished" podID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerID="acfca85b66fe60e68e3d514e25c02418fbb8d1f577c8563489417c05c7f14b48" exitCode=0 Oct 14 08:04:03 crc kubenswrapper[5002]: I1014 08:04:03.613480 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" event={"ID":"60f12264-cedf-4b85-8aa3-2707d34e6a00","Type":"ContainerDied","Data":"acfca85b66fe60e68e3d514e25c02418fbb8d1f577c8563489417c05c7f14b48"} Oct 14 08:04:03 crc kubenswrapper[5002]: I1014 08:04:03.613728 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rscf5" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="registry-server" containerID="cri-o://0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af" gracePeriod=2 Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.615997 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.624315 5002 generic.go:334] "Generic (PLEG): container finished" podID="e723db1a-1996-415e-a42e-e497689c4194" containerID="0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af" exitCode=0 Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.624371 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rscf5" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.624395 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerDied","Data":"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af"} Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.624425 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rscf5" event={"ID":"e723db1a-1996-415e-a42e-e497689c4194","Type":"ContainerDied","Data":"fdd9a790ab4c211c99f4c7bb0b79645ad2cee3bbbc2551da91e655d05b6d936f"} Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.624448 5002 scope.go:117] "RemoveContainer" containerID="0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.628448 5002 generic.go:334] "Generic (PLEG): container finished" podID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerID="6fadf81768c6b2fae56ab1294d6f63706de8311660f8ca26822752b0dacc5058" exitCode=0 Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.628479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" event={"ID":"60f12264-cedf-4b85-8aa3-2707d34e6a00","Type":"ContainerDied","Data":"6fadf81768c6b2fae56ab1294d6f63706de8311660f8ca26822752b0dacc5058"} Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.691370 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities\") pod \"e723db1a-1996-415e-a42e-e497689c4194\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.691434 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v85vg\" (UniqueName: \"kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg\") pod \"e723db1a-1996-415e-a42e-e497689c4194\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.691539 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content\") pod \"e723db1a-1996-415e-a42e-e497689c4194\" (UID: \"e723db1a-1996-415e-a42e-e497689c4194\") " Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.691949 5002 scope.go:117] "RemoveContainer" containerID="90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.693475 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities" (OuterVolumeSpecName: "utilities") pod "e723db1a-1996-415e-a42e-e497689c4194" (UID: "e723db1a-1996-415e-a42e-e497689c4194"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.702252 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg" (OuterVolumeSpecName: "kube-api-access-v85vg") pod "e723db1a-1996-415e-a42e-e497689c4194" (UID: "e723db1a-1996-415e-a42e-e497689c4194"). InnerVolumeSpecName "kube-api-access-v85vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.757396 5002 scope.go:117] "RemoveContainer" containerID="9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.757832 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e723db1a-1996-415e-a42e-e497689c4194" (UID: "e723db1a-1996-415e-a42e-e497689c4194"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.774336 5002 scope.go:117] "RemoveContainer" containerID="0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af" Oct 14 08:04:04 crc kubenswrapper[5002]: E1014 08:04:04.774753 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af\": container with ID starting with 0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af not found: ID does not exist" containerID="0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.774794 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af"} err="failed to get container status \"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af\": rpc error: code = NotFound desc = could not find container \"0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af\": container with ID starting with 0d91d56089b5ae000b55696b450abe02f8cdfa3c54096334ac3c60417b2a23af not found: ID does not exist" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.774816 5002 scope.go:117] "RemoveContainer" containerID="90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2" Oct 14 08:04:04 crc kubenswrapper[5002]: E1014 08:04:04.775227 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2\": container with ID starting with 90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2 not found: ID does not exist" containerID="90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.775294 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2"} err="failed to get container status \"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2\": rpc error: code = NotFound desc = could not find container \"90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2\": container with ID starting with 90e6a586be81de5ac49548ed71795ce8d175ea11e1240525b1b8bb4bbc0177c2 not found: ID does not exist" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.775322 5002 scope.go:117] "RemoveContainer" containerID="9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067" Oct 14 08:04:04 crc kubenswrapper[5002]: E1014 08:04:04.775669 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067\": container with ID starting with 9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067 not found: ID does not exist" containerID="9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.775694 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067"} err="failed to get container status \"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067\": rpc error: code = NotFound desc = could not find container \"9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067\": container with ID starting with 9eece9059b2fdc5ad51c73e1792d08301f0f8bf0d36dcd80f27c709d94ada067 not found: ID does not exist" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.792890 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.792910 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v85vg\" (UniqueName: \"kubernetes.io/projected/e723db1a-1996-415e-a42e-e497689c4194-kube-api-access-v85vg\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.792934 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e723db1a-1996-415e-a42e-e497689c4194-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.964034 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:04:04 crc kubenswrapper[5002]: I1014 08:04:04.968528 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rscf5"] Oct 14 08:04:05 crc kubenswrapper[5002]: I1014 08:04:05.737773 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e723db1a-1996-415e-a42e-e497689c4194" path="/var/lib/kubelet/pods/e723db1a-1996-415e-a42e-e497689c4194/volumes" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.145740 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.215556 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util\") pod \"60f12264-cedf-4b85-8aa3-2707d34e6a00\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.215684 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle\") pod \"60f12264-cedf-4b85-8aa3-2707d34e6a00\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.215812 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzqfp\" (UniqueName: \"kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp\") pod \"60f12264-cedf-4b85-8aa3-2707d34e6a00\" (UID: \"60f12264-cedf-4b85-8aa3-2707d34e6a00\") " Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.216540 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle" (OuterVolumeSpecName: "bundle") pod "60f12264-cedf-4b85-8aa3-2707d34e6a00" (UID: "60f12264-cedf-4b85-8aa3-2707d34e6a00"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.221231 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp" (OuterVolumeSpecName: "kube-api-access-dzqfp") pod "60f12264-cedf-4b85-8aa3-2707d34e6a00" (UID: "60f12264-cedf-4b85-8aa3-2707d34e6a00"). InnerVolumeSpecName "kube-api-access-dzqfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.317950 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzqfp\" (UniqueName: \"kubernetes.io/projected/60f12264-cedf-4b85-8aa3-2707d34e6a00-kube-api-access-dzqfp\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.317999 5002 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.482879 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util" (OuterVolumeSpecName: "util") pod "60f12264-cedf-4b85-8aa3-2707d34e6a00" (UID: "60f12264-cedf-4b85-8aa3-2707d34e6a00"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.521232 5002 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/60f12264-cedf-4b85-8aa3-2707d34e6a00-util\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.651391 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" event={"ID":"60f12264-cedf-4b85-8aa3-2707d34e6a00","Type":"ContainerDied","Data":"2adc7957c1a7ca5520235d689d1f16511fa7ec3a12a4e0f628042814d889fef8"} Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.651460 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2adc7957c1a7ca5520235d689d1f16511fa7ec3a12a4e0f628042814d889fef8" Oct 14 08:04:06 crc kubenswrapper[5002]: I1014 08:04:06.651479 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.357277 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358057 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="pull" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358074 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="pull" Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358109 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="extract" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358116 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="extract" Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358133 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="extract-utilities" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358142 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="extract-utilities" Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358161 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="registry-server" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358169 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="registry-server" Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358185 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="util" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358194 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="util" Oct 14 08:04:12 crc kubenswrapper[5002]: E1014 08:04:12.358220 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="extract-content" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358234 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="extract-content" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358565 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e723db1a-1996-415e-a42e-e497689c4194" containerName="registry-server" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.358579 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="60f12264-cedf-4b85-8aa3-2707d34e6a00" containerName="extract" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.360486 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.373580 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.502813 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z62tj\" (UniqueName: \"kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.503048 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.503102 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.518556 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.520175 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.528554 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ktjhn" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.551874 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.604608 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z62tj\" (UniqueName: \"kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.604677 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9stnv\" (UniqueName: \"kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv\") pod \"openstack-operator-controller-operator-64895cd698-k6rlc\" (UID: \"78216670-be87-44b5-be7e-b3e98f8e85e7\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.604719 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.604740 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.605156 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.605397 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.627496 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z62tj\" (UniqueName: \"kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj\") pod \"certified-operators-xhpvm\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.693475 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.706326 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9stnv\" (UniqueName: \"kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv\") pod \"openstack-operator-controller-operator-64895cd698-k6rlc\" (UID: \"78216670-be87-44b5-be7e-b3e98f8e85e7\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.727323 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9stnv\" (UniqueName: \"kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv\") pod \"openstack-operator-controller-operator-64895cd698-k6rlc\" (UID: \"78216670-be87-44b5-be7e-b3e98f8e85e7\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:12 crc kubenswrapper[5002]: I1014 08:04:12.842979 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.152873 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.283122 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:04:13 crc kubenswrapper[5002]: W1014 08:04:13.295151 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78216670_be87_44b5_be7e_b3e98f8e85e7.slice/crio-7bbe82ee5a8b29d536ed728ee1ba55a62eb6d5027cde5ab6565af2260fcc7dee WatchSource:0}: Error finding container 7bbe82ee5a8b29d536ed728ee1ba55a62eb6d5027cde5ab6565af2260fcc7dee: Status 404 returned error can't find the container with id 7bbe82ee5a8b29d536ed728ee1ba55a62eb6d5027cde5ab6565af2260fcc7dee Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.703153 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerStarted","Data":"7bbe82ee5a8b29d536ed728ee1ba55a62eb6d5027cde5ab6565af2260fcc7dee"} Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.704895 5002 generic.go:334] "Generic (PLEG): container finished" podID="702c5946-dd36-47ba-ab08-c65c014bd328" containerID="7ace40f0ac75913b67c1d68e3025499f20f29e6cbda9aec39d68d6bb14eb64f5" exitCode=0 Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.704934 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerDied","Data":"7ace40f0ac75913b67c1d68e3025499f20f29e6cbda9aec39d68d6bb14eb64f5"} Oct 14 08:04:13 crc kubenswrapper[5002]: I1014 08:04:13.704960 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerStarted","Data":"5a0f5eee88ad2da576e0c97bf355cbd1d3019da2a742773aec5c6191624b8136"} Oct 14 08:04:14 crc kubenswrapper[5002]: I1014 08:04:14.715585 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerStarted","Data":"c84a536a1732ec8dfed8c20fae4f19d45994be0b46f960747314c1e7dea31889"} Oct 14 08:04:15 crc kubenswrapper[5002]: I1014 08:04:15.725943 5002 generic.go:334] "Generic (PLEG): container finished" podID="702c5946-dd36-47ba-ab08-c65c014bd328" containerID="c84a536a1732ec8dfed8c20fae4f19d45994be0b46f960747314c1e7dea31889" exitCode=0 Oct 14 08:04:15 crc kubenswrapper[5002]: I1014 08:04:15.728581 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerDied","Data":"c84a536a1732ec8dfed8c20fae4f19d45994be0b46f960747314c1e7dea31889"} Oct 14 08:04:15 crc kubenswrapper[5002]: I1014 08:04:15.728635 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerStarted","Data":"b0a5d9771f66d18814e702c538a025b3bc6257c5f176be4df32903233d1664e9"} Oct 14 08:04:15 crc kubenswrapper[5002]: I1014 08:04:15.746223 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xhpvm" podStartSLOduration=2.184948302 podStartE2EDuration="3.746204209s" podCreationTimestamp="2025-10-14 08:04:12 +0000 UTC" firstStartedPulling="2025-10-14 08:04:13.706258134 +0000 UTC m=+786.687497586" lastFinishedPulling="2025-10-14 08:04:15.267514041 +0000 UTC m=+788.248753493" observedRunningTime="2025-10-14 08:04:15.745728297 +0000 UTC m=+788.726967779" watchObservedRunningTime="2025-10-14 08:04:15.746204209 +0000 UTC m=+788.727443671" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.153085 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.154174 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.166818 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.265388 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnrs6\" (UniqueName: \"kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.265446 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.265529 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.366718 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnrs6\" (UniqueName: \"kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.366776 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.366884 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.367174 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.367254 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.386298 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnrs6\" (UniqueName: \"kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6\") pod \"redhat-marketplace-5t2h6\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:16 crc kubenswrapper[5002]: I1014 08:04:16.484236 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:18 crc kubenswrapper[5002]: I1014 08:04:18.481320 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:18 crc kubenswrapper[5002]: W1014 08:04:18.487105 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba85522b_5a84_4822_bf3c_47d17bea2984.slice/crio-d83494ba94f00e3173c86a9a1a4440c69a1f4e610274251ca575097291cf2b8a WatchSource:0}: Error finding container d83494ba94f00e3173c86a9a1a4440c69a1f4e610274251ca575097291cf2b8a: Status 404 returned error can't find the container with id d83494ba94f00e3173c86a9a1a4440c69a1f4e610274251ca575097291cf2b8a Oct 14 08:04:18 crc kubenswrapper[5002]: I1014 08:04:18.742051 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerStarted","Data":"4843b6ff5cad6ca4bd2edfa89e75ba3f55cf28f5ad0b97de33d73fc582ecb619"} Oct 14 08:04:18 crc kubenswrapper[5002]: I1014 08:04:18.743916 5002 generic.go:334] "Generic (PLEG): container finished" podID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerID="011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf" exitCode=0 Oct 14 08:04:18 crc kubenswrapper[5002]: I1014 08:04:18.743970 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerDied","Data":"011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf"} Oct 14 08:04:18 crc kubenswrapper[5002]: I1014 08:04:18.743990 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerStarted","Data":"d83494ba94f00e3173c86a9a1a4440c69a1f4e610274251ca575097291cf2b8a"} Oct 14 08:04:21 crc kubenswrapper[5002]: I1014 08:04:21.765069 5002 generic.go:334] "Generic (PLEG): container finished" podID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerID="bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a" exitCode=0 Oct 14 08:04:21 crc kubenswrapper[5002]: I1014 08:04:21.765154 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerDied","Data":"bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a"} Oct 14 08:04:21 crc kubenswrapper[5002]: I1014 08:04:21.768269 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerStarted","Data":"9352e0d127ce543c017c1156b3d81e8c54c20be48de0865c50640db9feb31632"} Oct 14 08:04:21 crc kubenswrapper[5002]: I1014 08:04:21.768561 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:21 crc kubenswrapper[5002]: I1014 08:04:21.828026 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" podStartSLOduration=2.260807242 podStartE2EDuration="9.827998504s" podCreationTimestamp="2025-10-14 08:04:12 +0000 UTC" firstStartedPulling="2025-10-14 08:04:13.296723676 +0000 UTC m=+786.277963128" lastFinishedPulling="2025-10-14 08:04:20.863914928 +0000 UTC m=+793.845154390" observedRunningTime="2025-10-14 08:04:21.82295952 +0000 UTC m=+794.804199002" watchObservedRunningTime="2025-10-14 08:04:21.827998504 +0000 UTC m=+794.809237996" Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.693914 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.694241 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.778550 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerStarted","Data":"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad"} Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.783549 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.806783 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5t2h6" podStartSLOduration=3.308076715 podStartE2EDuration="6.80676631s" podCreationTimestamp="2025-10-14 08:04:16 +0000 UTC" firstStartedPulling="2025-10-14 08:04:18.746915012 +0000 UTC m=+791.728154464" lastFinishedPulling="2025-10-14 08:04:22.245604597 +0000 UTC m=+795.226844059" observedRunningTime="2025-10-14 08:04:22.802445295 +0000 UTC m=+795.783684757" watchObservedRunningTime="2025-10-14 08:04:22.80676631 +0000 UTC m=+795.788005772" Oct 14 08:04:22 crc kubenswrapper[5002]: I1014 08:04:22.841223 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:26 crc kubenswrapper[5002]: I1014 08:04:26.485011 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:26 crc kubenswrapper[5002]: I1014 08:04:26.485491 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:26 crc kubenswrapper[5002]: I1014 08:04:26.557634 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:27 crc kubenswrapper[5002]: I1014 08:04:27.878752 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.554204 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.555168 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xhpvm" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="registry-server" containerID="cri-o://b0a5d9771f66d18814e702c538a025b3bc6257c5f176be4df32903233d1664e9" gracePeriod=2 Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.826565 5002 generic.go:334] "Generic (PLEG): container finished" podID="702c5946-dd36-47ba-ab08-c65c014bd328" containerID="b0a5d9771f66d18814e702c538a025b3bc6257c5f176be4df32903233d1664e9" exitCode=0 Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.826763 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerDied","Data":"b0a5d9771f66d18814e702c538a025b3bc6257c5f176be4df32903233d1664e9"} Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.950692 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:28 crc kubenswrapper[5002]: I1014 08:04:28.989814 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.159441 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z62tj\" (UniqueName: \"kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj\") pod \"702c5946-dd36-47ba-ab08-c65c014bd328\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.159553 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content\") pod \"702c5946-dd36-47ba-ab08-c65c014bd328\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.159614 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities\") pod \"702c5946-dd36-47ba-ab08-c65c014bd328\" (UID: \"702c5946-dd36-47ba-ab08-c65c014bd328\") " Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.161613 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities" (OuterVolumeSpecName: "utilities") pod "702c5946-dd36-47ba-ab08-c65c014bd328" (UID: "702c5946-dd36-47ba-ab08-c65c014bd328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.172100 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj" (OuterVolumeSpecName: "kube-api-access-z62tj") pod "702c5946-dd36-47ba-ab08-c65c014bd328" (UID: "702c5946-dd36-47ba-ab08-c65c014bd328"). InnerVolumeSpecName "kube-api-access-z62tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.233174 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "702c5946-dd36-47ba-ab08-c65c014bd328" (UID: "702c5946-dd36-47ba-ab08-c65c014bd328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.262068 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z62tj\" (UniqueName: \"kubernetes.io/projected/702c5946-dd36-47ba-ab08-c65c014bd328-kube-api-access-z62tj\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.262120 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.262135 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702c5946-dd36-47ba-ab08-c65c014bd328-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.837470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xhpvm" event={"ID":"702c5946-dd36-47ba-ab08-c65c014bd328","Type":"ContainerDied","Data":"5a0f5eee88ad2da576e0c97bf355cbd1d3019da2a742773aec5c6191624b8136"} Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.837524 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xhpvm" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.837549 5002 scope.go:117] "RemoveContainer" containerID="b0a5d9771f66d18814e702c538a025b3bc6257c5f176be4df32903233d1664e9" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.837592 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5t2h6" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="registry-server" containerID="cri-o://8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad" gracePeriod=2 Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.866241 5002 scope.go:117] "RemoveContainer" containerID="c84a536a1732ec8dfed8c20fae4f19d45994be0b46f960747314c1e7dea31889" Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.869377 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.876599 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xhpvm"] Oct 14 08:04:29 crc kubenswrapper[5002]: I1014 08:04:29.888548 5002 scope.go:117] "RemoveContainer" containerID="7ace40f0ac75913b67c1d68e3025499f20f29e6cbda9aec39d68d6bb14eb64f5" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.329224 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.477377 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content\") pod \"ba85522b-5a84-4822-bf3c-47d17bea2984\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.477533 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnrs6\" (UniqueName: \"kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6\") pod \"ba85522b-5a84-4822-bf3c-47d17bea2984\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.477650 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities\") pod \"ba85522b-5a84-4822-bf3c-47d17bea2984\" (UID: \"ba85522b-5a84-4822-bf3c-47d17bea2984\") " Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.479522 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities" (OuterVolumeSpecName: "utilities") pod "ba85522b-5a84-4822-bf3c-47d17bea2984" (UID: "ba85522b-5a84-4822-bf3c-47d17bea2984"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.493169 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6" (OuterVolumeSpecName: "kube-api-access-fnrs6") pod "ba85522b-5a84-4822-bf3c-47d17bea2984" (UID: "ba85522b-5a84-4822-bf3c-47d17bea2984"). InnerVolumeSpecName "kube-api-access-fnrs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.505959 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba85522b-5a84-4822-bf3c-47d17bea2984" (UID: "ba85522b-5a84-4822-bf3c-47d17bea2984"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.580581 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnrs6\" (UniqueName: \"kubernetes.io/projected/ba85522b-5a84-4822-bf3c-47d17bea2984-kube-api-access-fnrs6\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.580632 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.580653 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba85522b-5a84-4822-bf3c-47d17bea2984-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.855796 5002 generic.go:334] "Generic (PLEG): container finished" podID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerID="8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad" exitCode=0 Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.855949 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerDied","Data":"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad"} Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.855997 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5t2h6" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.856066 5002 scope.go:117] "RemoveContainer" containerID="8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.856039 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5t2h6" event={"ID":"ba85522b-5a84-4822-bf3c-47d17bea2984","Type":"ContainerDied","Data":"d83494ba94f00e3173c86a9a1a4440c69a1f4e610274251ca575097291cf2b8a"} Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.888087 5002 scope.go:117] "RemoveContainer" containerID="bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.925071 5002 scope.go:117] "RemoveContainer" containerID="011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.959513 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.996027 5002 scope.go:117] "RemoveContainer" containerID="8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.996639 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5t2h6"] Oct 14 08:04:30 crc kubenswrapper[5002]: E1014 08:04:30.997167 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad\": container with ID starting with 8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad not found: ID does not exist" containerID="8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.997215 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad"} err="failed to get container status \"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad\": rpc error: code = NotFound desc = could not find container \"8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad\": container with ID starting with 8a693c0050e024b1593a56bcc4828ad10844e49d2899d64af128f07e4a1309ad not found: ID does not exist" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.997244 5002 scope.go:117] "RemoveContainer" containerID="bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a" Oct 14 08:04:30 crc kubenswrapper[5002]: E1014 08:04:30.997527 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a\": container with ID starting with bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a not found: ID does not exist" containerID="bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.997562 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a"} err="failed to get container status \"bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a\": rpc error: code = NotFound desc = could not find container \"bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a\": container with ID starting with bdf3ea92463e5812d20dd334a6c70b7d330ca3b77e7bc4c3f753a9261a8d3a6a not found: ID does not exist" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.997583 5002 scope.go:117] "RemoveContainer" containerID="011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf" Oct 14 08:04:30 crc kubenswrapper[5002]: E1014 08:04:30.997898 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf\": container with ID starting with 011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf not found: ID does not exist" containerID="011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf" Oct 14 08:04:30 crc kubenswrapper[5002]: I1014 08:04:30.998010 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf"} err="failed to get container status \"011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf\": rpc error: code = NotFound desc = could not find container \"011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf\": container with ID starting with 011d9e6b5a177de77c772627f8af2ec86d222d835559a3860309962a231f8bdf not found: ID does not exist" Oct 14 08:04:31 crc kubenswrapper[5002]: I1014 08:04:31.729265 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" path="/var/lib/kubelet/pods/702c5946-dd36-47ba-ab08-c65c014bd328/volumes" Oct 14 08:04:31 crc kubenswrapper[5002]: I1014 08:04:31.730820 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" path="/var/lib/kubelet/pods/ba85522b-5a84-4822-bf3c-47d17bea2984/volumes" Oct 14 08:04:32 crc kubenswrapper[5002]: I1014 08:04:32.847688 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.082732 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7"] Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083365 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="extract-utilities" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083377 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="extract-utilities" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083389 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083395 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083404 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="extract-content" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083410 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="extract-content" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083430 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="extract-content" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083435 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="extract-content" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083445 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="extract-utilities" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083450 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="extract-utilities" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.083460 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083466 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083570 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba85522b-5a84-4822-bf3c-47d17bea2984" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.083584 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="702c5946-dd36-47ba-ab08-c65c014bd328" containerName="registry-server" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.084170 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.087349 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zxxkn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.099857 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.109446 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.110355 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.115513 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-h2m48" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.126562 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.127694 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.136557 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-w8fc9" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.141770 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.165383 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.194292 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.195193 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.201172 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9dclz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.207964 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.232684 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbf4s\" (UniqueName: \"kubernetes.io/projected/578becbc-f063-4b98-b5a3-eb0656b24850-kube-api-access-cbf4s\") pod \"designate-operator-controller-manager-85d5d9dd78-9kfv7\" (UID: \"578becbc-f063-4b98-b5a3-eb0656b24850\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.232764 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxnhn\" (UniqueName: \"kubernetes.io/projected/cfe90f21-7b6a-4364-8c33-92b2880e9e0b-kube-api-access-nxnhn\") pod \"cinder-operator-controller-manager-7b7fb68549-lr6gb\" (UID: \"cfe90f21-7b6a-4364-8c33-92b2880e9e0b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.233065 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpxg6\" (UniqueName: \"kubernetes.io/projected/46630983-31be-46df-9e05-0506abbaf29e-kube-api-access-dpxg6\") pod \"barbican-operator-controller-manager-658bdf4b74-qpnq7\" (UID: \"46630983-31be-46df-9e05-0506abbaf29e\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.240899 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.244721 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.291213 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5pvs9" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.291373 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.331541 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.332561 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.344056 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.345066 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.346954 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-cktjf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347397 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbf4s\" (UniqueName: \"kubernetes.io/projected/578becbc-f063-4b98-b5a3-eb0656b24850-kube-api-access-cbf4s\") pod \"designate-operator-controller-manager-85d5d9dd78-9kfv7\" (UID: \"578becbc-f063-4b98-b5a3-eb0656b24850\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347442 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxnhn\" (UniqueName: \"kubernetes.io/projected/cfe90f21-7b6a-4364-8c33-92b2880e9e0b-kube-api-access-nxnhn\") pod \"cinder-operator-controller-manager-7b7fb68549-lr6gb\" (UID: \"cfe90f21-7b6a-4364-8c33-92b2880e9e0b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347478 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347483 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8bdn\" (UniqueName: \"kubernetes.io/projected/63c73302-dc00-4dfe-9d3f-d80cb181ffc2-kube-api-access-v8bdn\") pod \"heat-operator-controller-manager-858f76bbdd-5t4hm\" (UID: \"63c73302-dc00-4dfe-9d3f-d80cb181ffc2\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnqvl\" (UniqueName: \"kubernetes.io/projected/de07a029-a377-47c8-be2f-a31148f4948a-kube-api-access-vnqvl\") pod \"glance-operator-controller-manager-84b9b84486-8j2qn\" (UID: \"de07a029-a377-47c8-be2f-a31148f4948a\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347525 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347556 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kjg\" (UniqueName: \"kubernetes.io/projected/bbc859e5-324a-413c-a7ac-499a0b0b8fba-kube-api-access-m2kjg\") pod \"horizon-operator-controller-manager-7ffbcb7588-s6vkc\" (UID: \"bbc859e5-324a-413c-a7ac-499a0b0b8fba\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347582 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztmk8\" (UniqueName: \"kubernetes.io/projected/17c37ba2-ff55-4b60-a1db-01dec0522e2e-kube-api-access-ztmk8\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347596 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-k8pnq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.347629 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpxg6\" (UniqueName: \"kubernetes.io/projected/46630983-31be-46df-9e05-0506abbaf29e-kube-api-access-dpxg6\") pod \"barbican-operator-controller-manager-658bdf4b74-qpnq7\" (UID: \"46630983-31be-46df-9e05-0506abbaf29e\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.358901 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.363349 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.364434 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.367524 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tznpq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.374895 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.376041 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpxg6\" (UniqueName: \"kubernetes.io/projected/46630983-31be-46df-9e05-0506abbaf29e-kube-api-access-dpxg6\") pod \"barbican-operator-controller-manager-658bdf4b74-qpnq7\" (UID: \"46630983-31be-46df-9e05-0506abbaf29e\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.382921 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.389881 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.390202 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxnhn\" (UniqueName: \"kubernetes.io/projected/cfe90f21-7b6a-4364-8c33-92b2880e9e0b-kube-api-access-nxnhn\") pod \"cinder-operator-controller-manager-7b7fb68549-lr6gb\" (UID: \"cfe90f21-7b6a-4364-8c33-92b2880e9e0b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.390795 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.391496 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.391824 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.393883 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.395077 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-t4j22" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.395224 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-vk7bn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.401189 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.402358 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.408406 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbf4s\" (UniqueName: \"kubernetes.io/projected/578becbc-f063-4b98-b5a3-eb0656b24850-kube-api-access-cbf4s\") pod \"designate-operator-controller-manager-85d5d9dd78-9kfv7\" (UID: \"578becbc-f063-4b98-b5a3-eb0656b24850\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.408504 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.409452 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.410142 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.410440 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.413180 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-82c6h"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.417828 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.418183 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-m2kbk" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.418366 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-r2g57" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.418700 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.429564 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.442190 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-jdl7c" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.447043 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448336 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8bdn\" (UniqueName: \"kubernetes.io/projected/63c73302-dc00-4dfe-9d3f-d80cb181ffc2-kube-api-access-v8bdn\") pod \"heat-operator-controller-manager-858f76bbdd-5t4hm\" (UID: \"63c73302-dc00-4dfe-9d3f-d80cb181ffc2\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448373 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnqvl\" (UniqueName: \"kubernetes.io/projected/de07a029-a377-47c8-be2f-a31148f4948a-kube-api-access-vnqvl\") pod \"glance-operator-controller-manager-84b9b84486-8j2qn\" (UID: \"de07a029-a377-47c8-be2f-a31148f4948a\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448398 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k8wb\" (UniqueName: \"kubernetes.io/projected/9c72f60f-c311-45eb-84a1-8c95164be376-kube-api-access-7k8wb\") pod \"nova-operator-controller-manager-5df598886f-82c6h\" (UID: \"9c72f60f-c311-45eb-84a1-8c95164be376\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448418 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpdlk\" (UniqueName: \"kubernetes.io/projected/2041ec73-97de-43ea-ae4e-e8a1d7521157-kube-api-access-zpdlk\") pod \"manila-operator-controller-manager-5f67fbc655-59jzf\" (UID: \"2041ec73-97de-43ea-ae4e-e8a1d7521157\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448435 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448452 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kjg\" (UniqueName: \"kubernetes.io/projected/bbc859e5-324a-413c-a7ac-499a0b0b8fba-kube-api-access-m2kjg\") pod \"horizon-operator-controller-manager-7ffbcb7588-s6vkc\" (UID: \"bbc859e5-324a-413c-a7ac-499a0b0b8fba\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448471 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcvk7\" (UniqueName: \"kubernetes.io/projected/57a0f853-8feb-4064-9ed2-3460f2f66901-kube-api-access-fcvk7\") pod \"mariadb-operator-controller-manager-f9fb45f8f-xsb2d\" (UID: \"57a0f853-8feb-4064-9ed2-3460f2f66901\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448496 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztmk8\" (UniqueName: \"kubernetes.io/projected/17c37ba2-ff55-4b60-a1db-01dec0522e2e-kube-api-access-ztmk8\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448517 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6ljw\" (UniqueName: \"kubernetes.io/projected/95e654ed-72c1-4f5b-b670-a89c96692455-kube-api-access-g6ljw\") pod \"neutron-operator-controller-manager-79d585cb66-69rk7\" (UID: \"95e654ed-72c1-4f5b-b670-a89c96692455\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448550 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x2zp\" (UniqueName: \"kubernetes.io/projected/4b405ac3-bbcd-4432-8f99-2862ef406541-kube-api-access-9x2zp\") pod \"keystone-operator-controller-manager-55b6b7c7b8-kqdfb\" (UID: \"4b405ac3-bbcd-4432-8f99-2862ef406541\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.448566 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqwl8\" (UniqueName: \"kubernetes.io/projected/e1488db8-3d22-4c1f-8400-8d292b48027a-kube-api-access-gqwl8\") pod \"ironic-operator-controller-manager-9c5c78d49-7jcbw\" (UID: \"e1488db8-3d22-4c1f-8400-8d292b48027a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.449078 5002 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.449118 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert podName:17c37ba2-ff55-4b60-a1db-01dec0522e2e nodeName:}" failed. No retries permitted until 2025-10-14 08:04:57.949103001 +0000 UTC m=+830.930342453 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert") pod "infra-operator-controller-manager-656bcbd775-d9ftz" (UID: "17c37ba2-ff55-4b60-a1db-01dec0522e2e") : secret "infra-operator-webhook-server-cert" not found Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.453999 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.458672 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.458770 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.470957 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8ngfl" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.474286 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-82c6h"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.480961 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztmk8\" (UniqueName: \"kubernetes.io/projected/17c37ba2-ff55-4b60-a1db-01dec0522e2e-kube-api-access-ztmk8\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.481570 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.486337 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.487499 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.489816 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8bdn\" (UniqueName: \"kubernetes.io/projected/63c73302-dc00-4dfe-9d3f-d80cb181ffc2-kube-api-access-v8bdn\") pod \"heat-operator-controller-manager-858f76bbdd-5t4hm\" (UID: \"63c73302-dc00-4dfe-9d3f-d80cb181ffc2\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.501386 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnqvl\" (UniqueName: \"kubernetes.io/projected/de07a029-a377-47c8-be2f-a31148f4948a-kube-api-access-vnqvl\") pod \"glance-operator-controller-manager-84b9b84486-8j2qn\" (UID: \"de07a029-a377-47c8-be2f-a31148f4948a\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.508277 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gfv6m" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.508438 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.520387 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.522262 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.523211 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.529141 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fck5g" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.529204 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.538690 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kjg\" (UniqueName: \"kubernetes.io/projected/bbc859e5-324a-413c-a7ac-499a0b0b8fba-kube-api-access-m2kjg\") pod \"horizon-operator-controller-manager-7ffbcb7588-s6vkc\" (UID: \"bbc859e5-324a-413c-a7ac-499a0b0b8fba\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.550459 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.551659 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.551884 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps8qh\" (UniqueName: \"kubernetes.io/projected/05103074-be63-4240-bd7e-737c3c6ff4d5-kube-api-access-ps8qh\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.551970 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k8wb\" (UniqueName: \"kubernetes.io/projected/9c72f60f-c311-45eb-84a1-8c95164be376-kube-api-access-7k8wb\") pod \"nova-operator-controller-manager-5df598886f-82c6h\" (UID: \"9c72f60f-c311-45eb-84a1-8c95164be376\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552006 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpdlk\" (UniqueName: \"kubernetes.io/projected/2041ec73-97de-43ea-ae4e-e8a1d7521157-kube-api-access-zpdlk\") pod \"manila-operator-controller-manager-5f67fbc655-59jzf\" (UID: \"2041ec73-97de-43ea-ae4e-e8a1d7521157\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552044 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcvk7\" (UniqueName: \"kubernetes.io/projected/57a0f853-8feb-4064-9ed2-3460f2f66901-kube-api-access-fcvk7\") pod \"mariadb-operator-controller-manager-f9fb45f8f-xsb2d\" (UID: \"57a0f853-8feb-4064-9ed2-3460f2f66901\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552089 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6ljw\" (UniqueName: \"kubernetes.io/projected/95e654ed-72c1-4f5b-b670-a89c96692455-kube-api-access-g6ljw\") pod \"neutron-operator-controller-manager-79d585cb66-69rk7\" (UID: \"95e654ed-72c1-4f5b-b670-a89c96692455\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552104 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj92v\" (UniqueName: \"kubernetes.io/projected/8e0847dd-54d1-44f7-8df9-b1378b00cb67-kube-api-access-qj92v\") pod \"ovn-operator-controller-manager-79df5fb58c-xj9vc\" (UID: \"8e0847dd-54d1-44f7-8df9-b1378b00cb67\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552136 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552158 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r4mr\" (UniqueName: \"kubernetes.io/projected/27e01872-38da-4f89-8232-c8edaec917f8-kube-api-access-8r4mr\") pod \"octavia-operator-controller-manager-69fdcfc5f5-cn7pq\" (UID: \"27e01872-38da-4f89-8232-c8edaec917f8\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552177 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x2zp\" (UniqueName: \"kubernetes.io/projected/4b405ac3-bbcd-4432-8f99-2862ef406541-kube-api-access-9x2zp\") pod \"keystone-operator-controller-manager-55b6b7c7b8-kqdfb\" (UID: \"4b405ac3-bbcd-4432-8f99-2862ef406541\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.552192 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqwl8\" (UniqueName: \"kubernetes.io/projected/e1488db8-3d22-4c1f-8400-8d292b48027a-kube-api-access-gqwl8\") pod \"ironic-operator-controller-manager-9c5c78d49-7jcbw\" (UID: \"e1488db8-3d22-4c1f-8400-8d292b48027a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.556935 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.558166 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-9zt2k" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.570946 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.571960 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.572653 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.573211 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.578539 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w5xvt" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.578732 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-675v4" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.589577 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcvk7\" (UniqueName: \"kubernetes.io/projected/57a0f853-8feb-4064-9ed2-3460f2f66901-kube-api-access-fcvk7\") pod \"mariadb-operator-controller-manager-f9fb45f8f-xsb2d\" (UID: \"57a0f853-8feb-4064-9ed2-3460f2f66901\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.590947 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.593093 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6ljw\" (UniqueName: \"kubernetes.io/projected/95e654ed-72c1-4f5b-b670-a89c96692455-kube-api-access-g6ljw\") pod \"neutron-operator-controller-manager-79d585cb66-69rk7\" (UID: \"95e654ed-72c1-4f5b-b670-a89c96692455\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.595337 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpdlk\" (UniqueName: \"kubernetes.io/projected/2041ec73-97de-43ea-ae4e-e8a1d7521157-kube-api-access-zpdlk\") pod \"manila-operator-controller-manager-5f67fbc655-59jzf\" (UID: \"2041ec73-97de-43ea-ae4e-e8a1d7521157\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.603353 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqwl8\" (UniqueName: \"kubernetes.io/projected/e1488db8-3d22-4c1f-8400-8d292b48027a-kube-api-access-gqwl8\") pod \"ironic-operator-controller-manager-9c5c78d49-7jcbw\" (UID: \"e1488db8-3d22-4c1f-8400-8d292b48027a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.608472 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.615511 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.621406 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.621823 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x2zp\" (UniqueName: \"kubernetes.io/projected/4b405ac3-bbcd-4432-8f99-2862ef406541-kube-api-access-9x2zp\") pod \"keystone-operator-controller-manager-55b6b7c7b8-kqdfb\" (UID: \"4b405ac3-bbcd-4432-8f99-2862ef406541\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.626233 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k8wb\" (UniqueName: \"kubernetes.io/projected/9c72f60f-c311-45eb-84a1-8c95164be376-kube-api-access-7k8wb\") pod \"nova-operator-controller-manager-5df598886f-82c6h\" (UID: \"9c72f60f-c311-45eb-84a1-8c95164be376\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.637675 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.638975 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.645039 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.646085 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.647642 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4zrlm" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.665921 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj92v\" (UniqueName: \"kubernetes.io/projected/8e0847dd-54d1-44f7-8df9-b1378b00cb67-kube-api-access-qj92v\") pod \"ovn-operator-controller-manager-79df5fb58c-xj9vc\" (UID: \"8e0847dd-54d1-44f7-8df9-b1378b00cb67\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.670757 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.670817 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r4mr\" (UniqueName: \"kubernetes.io/projected/27e01872-38da-4f89-8232-c8edaec917f8-kube-api-access-8r4mr\") pod \"octavia-operator-controller-manager-69fdcfc5f5-cn7pq\" (UID: \"27e01872-38da-4f89-8232-c8edaec917f8\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.671182 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps8qh\" (UniqueName: \"kubernetes.io/projected/05103074-be63-4240-bd7e-737c3c6ff4d5-kube-api-access-ps8qh\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.674613 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.675594 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.679588 5002 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 08:04:57 crc kubenswrapper[5002]: E1014 08:04:57.679667 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert podName:05103074-be63-4240-bd7e-737c3c6ff4d5 nodeName:}" failed. No retries permitted until 2025-10-14 08:04:58.179644456 +0000 UTC m=+831.160883908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert") pod "openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" (UID: "05103074-be63-4240-bd7e-737c3c6ff4d5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.709799 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps8qh\" (UniqueName: \"kubernetes.io/projected/05103074-be63-4240-bd7e-737c3c6ff4d5-kube-api-access-ps8qh\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.723613 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r4mr\" (UniqueName: \"kubernetes.io/projected/27e01872-38da-4f89-8232-c8edaec917f8-kube-api-access-8r4mr\") pod \"octavia-operator-controller-manager-69fdcfc5f5-cn7pq\" (UID: \"27e01872-38da-4f89-8232-c8edaec917f8\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.729873 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj92v\" (UniqueName: \"kubernetes.io/projected/8e0847dd-54d1-44f7-8df9-b1378b00cb67-kube-api-access-qj92v\") pod \"ovn-operator-controller-manager-79df5fb58c-xj9vc\" (UID: \"8e0847dd-54d1-44f7-8df9-b1378b00cb67\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.751067 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.777216 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j8j7\" (UniqueName: \"kubernetes.io/projected/b04b0d3d-03a0-4f12-814f-599f2899962b-kube-api-access-9j8j7\") pod \"swift-operator-controller-manager-db6d7f97b-74c8t\" (UID: \"b04b0d3d-03a0-4f12-814f-599f2899962b\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.777283 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhx88\" (UniqueName: \"kubernetes.io/projected/20fefb60-1453-4a27-9f7a-518df5b902fc-kube-api-access-nhx88\") pod \"placement-operator-controller-manager-68b6c87b68-tws79\" (UID: \"20fefb60-1453-4a27-9f7a-518df5b902fc\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.777367 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98fqt\" (UniqueName: \"kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt\") pod \"test-operator-controller-manager-5458f77c4-xd9hn\" (UID: \"e4e3eea0-46ac-4106-bf62-6cd798e2f96b\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.777429 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b997b\" (UniqueName: \"kubernetes.io/projected/42f32565-187e-4c71-a2c3-35eb988b8243-kube-api-access-b997b\") pod \"telemetry-operator-controller-manager-67cfc6749b-fp4rw\" (UID: \"42f32565-187e-4c71-a2c3-35eb988b8243\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.817051 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.822275 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.823216 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.823285 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.829658 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-spcvf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.834800 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.843538 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.860809 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.862228 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.867196 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-f26z4" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.867368 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.878507 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98fqt\" (UniqueName: \"kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt\") pod \"test-operator-controller-manager-5458f77c4-xd9hn\" (UID: \"e4e3eea0-46ac-4106-bf62-6cd798e2f96b\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.878778 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b997b\" (UniqueName: \"kubernetes.io/projected/42f32565-187e-4c71-a2c3-35eb988b8243-kube-api-access-b997b\") pod \"telemetry-operator-controller-manager-67cfc6749b-fp4rw\" (UID: \"42f32565-187e-4c71-a2c3-35eb988b8243\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.878887 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j8j7\" (UniqueName: \"kubernetes.io/projected/b04b0d3d-03a0-4f12-814f-599f2899962b-kube-api-access-9j8j7\") pod \"swift-operator-controller-manager-db6d7f97b-74c8t\" (UID: \"b04b0d3d-03a0-4f12-814f-599f2899962b\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.878987 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhx88\" (UniqueName: \"kubernetes.io/projected/20fefb60-1453-4a27-9f7a-518df5b902fc-kube-api-access-nhx88\") pod \"placement-operator-controller-manager-68b6c87b68-tws79\" (UID: \"20fefb60-1453-4a27-9f7a-518df5b902fc\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.881128 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.888370 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.889609 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.892399 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt"] Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.898441 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-s447x" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.898811 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.899073 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.921831 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j8j7\" (UniqueName: \"kubernetes.io/projected/b04b0d3d-03a0-4f12-814f-599f2899962b-kube-api-access-9j8j7\") pod \"swift-operator-controller-manager-db6d7f97b-74c8t\" (UID: \"b04b0d3d-03a0-4f12-814f-599f2899962b\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.923750 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b997b\" (UniqueName: \"kubernetes.io/projected/42f32565-187e-4c71-a2c3-35eb988b8243-kube-api-access-b997b\") pod \"telemetry-operator-controller-manager-67cfc6749b-fp4rw\" (UID: \"42f32565-187e-4c71-a2c3-35eb988b8243\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.925420 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhx88\" (UniqueName: \"kubernetes.io/projected/20fefb60-1453-4a27-9f7a-518df5b902fc-kube-api-access-nhx88\") pod \"placement-operator-controller-manager-68b6c87b68-tws79\" (UID: \"20fefb60-1453-4a27-9f7a-518df5b902fc\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.928737 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98fqt\" (UniqueName: \"kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt\") pod \"test-operator-controller-manager-5458f77c4-xd9hn\" (UID: \"e4e3eea0-46ac-4106-bf62-6cd798e2f96b\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.981442 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8848l\" (UniqueName: \"kubernetes.io/projected/516be58e-c541-4fbe-b4b8-127e07a0b8b7-kube-api-access-8848l\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.981987 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.982031 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpnbx\" (UniqueName: \"kubernetes.io/projected/74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8-kube-api-access-wpnbx\") pod \"watcher-operator-controller-manager-7f554bff7b-lwlwj\" (UID: \"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.982083 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqj9l\" (UniqueName: \"kubernetes.io/projected/b1921718-5347-4bfd-a327-9c38671b4888-kube-api-access-cqj9l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s89dt\" (UID: \"b1921718-5347-4bfd-a327-9c38671b4888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.982155 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.986199 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/17c37ba2-ff55-4b60-a1db-01dec0522e2e-cert\") pod \"infra-operator-controller-manager-656bcbd775-d9ftz\" (UID: \"17c37ba2-ff55-4b60-a1db-01dec0522e2e\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:57 crc kubenswrapper[5002]: I1014 08:04:57.986424 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.048379 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.088689 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.091171 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpnbx\" (UniqueName: \"kubernetes.io/projected/74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8-kube-api-access-wpnbx\") pod \"watcher-operator-controller-manager-7f554bff7b-lwlwj\" (UID: \"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.091205 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqj9l\" (UniqueName: \"kubernetes.io/projected/b1921718-5347-4bfd-a327-9c38671b4888-kube-api-access-cqj9l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s89dt\" (UID: \"b1921718-5347-4bfd-a327-9c38671b4888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.091237 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.091303 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8848l\" (UniqueName: \"kubernetes.io/projected/516be58e-c541-4fbe-b4b8-127e07a0b8b7-kube-api-access-8848l\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:58 crc kubenswrapper[5002]: E1014 08:04:58.091824 5002 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 08:04:58 crc kubenswrapper[5002]: E1014 08:04:58.091881 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert podName:516be58e-c541-4fbe-b4b8-127e07a0b8b7 nodeName:}" failed. No retries permitted until 2025-10-14 08:04:58.591869155 +0000 UTC m=+831.573108607 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert") pod "openstack-operator-controller-manager-7fb8c88b76-k6scp" (UID: "516be58e-c541-4fbe-b4b8-127e07a0b8b7") : secret "webhook-server-cert" not found Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.121276 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.123042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8848l\" (UniqueName: \"kubernetes.io/projected/516be58e-c541-4fbe-b4b8-127e07a0b8b7-kube-api-access-8848l\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.136002 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpnbx\" (UniqueName: \"kubernetes.io/projected/74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8-kube-api-access-wpnbx\") pod \"watcher-operator-controller-manager-7f554bff7b-lwlwj\" (UID: \"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.142156 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqj9l\" (UniqueName: \"kubernetes.io/projected/b1921718-5347-4bfd-a327-9c38671b4888-kube-api-access-cqj9l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-s89dt\" (UID: \"b1921718-5347-4bfd-a327-9c38671b4888\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.168716 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.181307 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.193166 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:58 crc kubenswrapper[5002]: E1014 08:04:58.193899 5002 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 08:04:58 crc kubenswrapper[5002]: E1014 08:04:58.194029 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert podName:05103074-be63-4240-bd7e-737c3c6ff4d5 nodeName:}" failed. No retries permitted until 2025-10-14 08:04:59.194013424 +0000 UTC m=+832.175252876 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert") pod "openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" (UID: "05103074-be63-4240-bd7e-737c3c6ff4d5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.231645 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.353178 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.379389 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.384508 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.602051 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.605969 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/516be58e-c541-4fbe-b4b8-127e07a0b8b7-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-k6scp\" (UID: \"516be58e-c541-4fbe-b4b8-127e07a0b8b7\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.722187 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.742803 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.747126 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc"] Oct 14 08:04:58 crc kubenswrapper[5002]: I1014 08:04:58.822351 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.091206 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" event={"ID":"63c73302-dc00-4dfe-9d3f-d80cb181ffc2","Type":"ContainerStarted","Data":"8689692be83f158c6c08a2269180e97b6296c35c8bfbf26d0692c39707059d3c"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.093984 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" event={"ID":"578becbc-f063-4b98-b5a3-eb0656b24850","Type":"ContainerStarted","Data":"b2de7baf7a75ae4780312e5575082c633cce3fcc232e434c27d5f939458b1fc9"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.094961 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" event={"ID":"cfe90f21-7b6a-4364-8c33-92b2880e9e0b","Type":"ContainerStarted","Data":"501c2a622f864b099bc99ca70f6905ccd909b04f47ba517e4c68c5e40ccfac07"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.095975 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" event={"ID":"46630983-31be-46df-9e05-0506abbaf29e","Type":"ContainerStarted","Data":"fae2adf8a8efb26e79afd1b797624a9b1fd3246cc81227b3a4175c99c382b9cf"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.097108 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" event={"ID":"de07a029-a377-47c8-be2f-a31148f4948a","Type":"ContainerStarted","Data":"b76ccc2b4743aac1c569101637e35a470347b16d8d663dc9f9ba9f3ce60012c5"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.098487 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" event={"ID":"8e0847dd-54d1-44f7-8df9-b1378b00cb67","Type":"ContainerStarted","Data":"1b9bfb2a29914fddec896862735794f173c67cef8355684051f488c422d1b64c"} Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.115354 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.145510 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.168287 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.181923 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.199941 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq"] Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.208432 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27e01872_38da_4f89_8232_c8edaec917f8.slice/crio-abc8b512e2f808d00088466b40832199b7e210a574314f9bb5baafe966eb7fb4 WatchSource:0}: Error finding container abc8b512e2f808d00088466b40832199b7e210a574314f9bb5baafe966eb7fb4: Status 404 returned error can't find the container with id abc8b512e2f808d00088466b40832199b7e210a574314f9bb5baafe966eb7fb4 Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.217256 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqwl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-9c5c78d49-7jcbw_openstack-operators(e1488db8-3d22-4c1f-8400-8d292b48027a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.221037 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.224707 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.226613 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/05103074-be63-4240-bd7e-737c3c6ff4d5-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d44848n4xg8\" (UID: \"05103074-be63-4240-bd7e-737c3c6ff4d5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.231850 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t"] Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.245114 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9j8j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-db6d7f97b-74c8t_openstack-operators(b04b0d3d-03a0-4f12-814f-599f2899962b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.245227 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b997b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-67cfc6749b-fp4rw_openstack-operators(42f32565-187e-4c71-a2c3-35eb988b8243): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.248746 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1921718_5347_4bfd_a327_9c38671b4888.slice/crio-3a1d8479d43ea0e67d878664e349fd0084e2b27d8b2f8e3d93356167f0b00be5 WatchSource:0}: Error finding container 3a1d8479d43ea0e67d878664e349fd0084e2b27d8b2f8e3d93356167f0b00be5: Status 404 returned error can't find the container with id 3a1d8479d43ea0e67d878664e349fd0084e2b27d8b2f8e3d93356167f0b00be5 Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.249146 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-98fqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5458f77c4-xd9hn_openstack-operators(e4e3eea0-46ac-4106-bf62-6cd798e2f96b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.257887 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-82c6h"] Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.259979 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20fefb60_1453_4a27_9f7a_518df5b902fc.slice/crio-8e5b8c392d4ead4494acb8c9b3378e4a4cbd69f9fbe4171dd1701e44160c3b00 WatchSource:0}: Error finding container 8e5b8c392d4ead4494acb8c9b3378e4a4cbd69f9fbe4171dd1701e44160c3b00: Status 404 returned error can't find the container with id 8e5b8c392d4ead4494acb8c9b3378e4a4cbd69f9fbe4171dd1701e44160c3b00 Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.272944 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17c37ba2_ff55_4b60_a1db_01dec0522e2e.slice/crio-5108ce05aea742f8b155eb2136f1b2400956040fa7d2426da2e96d24faedf61b WatchSource:0}: Error finding container 5108ce05aea742f8b155eb2136f1b2400956040fa7d2426da2e96d24faedf61b: Status 404 returned error can't find the container with id 5108ce05aea742f8b155eb2136f1b2400956040fa7d2426da2e96d24faedf61b Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.279993 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nhx88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-tws79_openstack-operators(20fefb60-1453-4a27-9f7a-518df5b902fc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.281370 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ztmk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-d9ftz_openstack-operators(17c37ba2-ff55-4b60-a1db-01dec0522e2e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.281421 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz"] Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.282024 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cqj9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-s89dt_openstack-operators(b1921718-5347-4bfd-a327-9c38671b4888): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.283435 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" podUID="b1921718-5347-4bfd-a327-9c38671b4888" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.304014 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.324898 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.336100 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.345039 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.355752 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.368738 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.382818 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj"] Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.431947 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp"] Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.497239 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod516be58e_c541_4fbe_b4b8_127e07a0b8b7.slice/crio-893c15d28b2d2db3aa073b57189f438203cbc6a0db69628cc13b84069787537e WatchSource:0}: Error finding container 893c15d28b2d2db3aa073b57189f438203cbc6a0db69628cc13b84069787537e: Status 404 returned error can't find the container with id 893c15d28b2d2db3aa073b57189f438203cbc6a0db69628cc13b84069787537e Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.647156 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" podUID="e1488db8-3d22-4c1f-8400-8d292b48027a" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.705253 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" podUID="b04b0d3d-03a0-4f12-814f-599f2899962b" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.722935 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.724660 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" podUID="42f32565-187e-4c71-a2c3-35eb988b8243" Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.766596 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" podUID="20fefb60-1453-4a27-9f7a-518df5b902fc" Oct 14 08:04:59 crc kubenswrapper[5002]: I1014 08:04:59.778933 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8"] Oct 14 08:04:59 crc kubenswrapper[5002]: E1014 08:04:59.810459 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" podUID="17c37ba2-ff55-4b60-a1db-01dec0522e2e" Oct 14 08:04:59 crc kubenswrapper[5002]: W1014 08:04:59.824935 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05103074_be63_4240_bd7e_737c3c6ff4d5.slice/crio-aaaa4aa0f3d27a3ef86c12f9fdb7154f3f27ae326d88f08fb7bf70d44e114aa8 WatchSource:0}: Error finding container aaaa4aa0f3d27a3ef86c12f9fdb7154f3f27ae326d88f08fb7bf70d44e114aa8: Status 404 returned error can't find the container with id aaaa4aa0f3d27a3ef86c12f9fdb7154f3f27ae326d88f08fb7bf70d44e114aa8 Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.026041 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.030856 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.043730 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.137470 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.137543 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6xfj\" (UniqueName: \"kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.137623 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.142598 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" event={"ID":"27e01872-38da-4f89-8232-c8edaec917f8","Type":"ContainerStarted","Data":"abc8b512e2f808d00088466b40832199b7e210a574314f9bb5baafe966eb7fb4"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.147747 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" event={"ID":"57a0f853-8feb-4064-9ed2-3460f2f66901","Type":"ContainerStarted","Data":"859b8906701c34620b92df0ec51eeef4685c2d1a0a7ff499bd501fbb1f612cde"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.172071 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" event={"ID":"516be58e-c541-4fbe-b4b8-127e07a0b8b7","Type":"ContainerStarted","Data":"701739d6da341532e3b329e4e434c43fe9aa9845f5fe811ed97e1d5b1c14226a"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.172124 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" event={"ID":"516be58e-c541-4fbe-b4b8-127e07a0b8b7","Type":"ContainerStarted","Data":"964d5dceb3271fc4cb5c94d80ac4a34f80182ea3ef9175d5621602e94b743337"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.172138 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" event={"ID":"516be58e-c541-4fbe-b4b8-127e07a0b8b7","Type":"ContainerStarted","Data":"893c15d28b2d2db3aa073b57189f438203cbc6a0db69628cc13b84069787537e"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.173047 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.201002 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" event={"ID":"b1921718-5347-4bfd-a327-9c38671b4888","Type":"ContainerStarted","Data":"3a1d8479d43ea0e67d878664e349fd0084e2b27d8b2f8e3d93356167f0b00be5"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.209223 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" podUID="b1921718-5347-4bfd-a327-9c38671b4888" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.218850 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" podStartSLOduration=3.218816396 podStartE2EDuration="3.218816396s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:05:00.209475017 +0000 UTC m=+833.190714479" watchObservedRunningTime="2025-10-14 08:05:00.218816396 +0000 UTC m=+833.200055848" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.221470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" event={"ID":"b04b0d3d-03a0-4f12-814f-599f2899962b","Type":"ContainerStarted","Data":"ffcd656270efa5632235990ac6df538d714511e678c833a56aeedb2ae215c0fd"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.221512 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" event={"ID":"b04b0d3d-03a0-4f12-814f-599f2899962b","Type":"ContainerStarted","Data":"a8e91ac3559a0393653589be66c1939fe176b6d2e0647bfc8c55544743856856"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.227003 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" podUID="b04b0d3d-03a0-4f12-814f-599f2899962b" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.227716 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" event={"ID":"2041ec73-97de-43ea-ae4e-e8a1d7521157","Type":"ContainerStarted","Data":"e212dcb3a981eb1b951087612bdca1dab76eeb2807dd5777170a7b48edb013ab"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.240416 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.240468 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.240504 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6xfj\" (UniqueName: \"kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.240550 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" event={"ID":"42f32565-187e-4c71-a2c3-35eb988b8243","Type":"ContainerStarted","Data":"3de1d75f711e9c5811a7c49ae35f99eed220111f9ebd0d550bc8c0e92b6dac74"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.240576 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" event={"ID":"42f32565-187e-4c71-a2c3-35eb988b8243","Type":"ContainerStarted","Data":"5a1ce0d49c8301a5b099522789a8e88affcb299ec2a09c67ab6bff001b745a35"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.241091 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.241314 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.242034 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" podUID="42f32565-187e-4c71-a2c3-35eb988b8243" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.244998 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" event={"ID":"05103074-be63-4240-bd7e-737c3c6ff4d5","Type":"ContainerStarted","Data":"aaaa4aa0f3d27a3ef86c12f9fdb7154f3f27ae326d88f08fb7bf70d44e114aa8"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.246460 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" event={"ID":"20fefb60-1453-4a27-9f7a-518df5b902fc","Type":"ContainerStarted","Data":"bb0346b7236be4480683657e10502e0480d703ab72d7841b1e32e2a120db700a"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.246547 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" event={"ID":"20fefb60-1453-4a27-9f7a-518df5b902fc","Type":"ContainerStarted","Data":"8e5b8c392d4ead4494acb8c9b3378e4a4cbd69f9fbe4171dd1701e44160c3b00"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.249067 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" podUID="20fefb60-1453-4a27-9f7a-518df5b902fc" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.249531 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" event={"ID":"bbc859e5-324a-413c-a7ac-499a0b0b8fba","Type":"ContainerStarted","Data":"c07d844f8200dc9717f23c81104126484ffd1ed207f5e294572b5c668dc8e816"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.252365 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" event={"ID":"4b405ac3-bbcd-4432-8f99-2862ef406541","Type":"ContainerStarted","Data":"3fa2c4526e5a248c7a158a79c29a49dd1d0c842cc036b0d745ccd290588978ea"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.271526 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" event={"ID":"e1488db8-3d22-4c1f-8400-8d292b48027a","Type":"ContainerStarted","Data":"6ed812126e9a5a74b1212594b58f78c987a7dff7808312aceed79dfb5d6d743b"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.271568 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" event={"ID":"e1488db8-3d22-4c1f-8400-8d292b48027a","Type":"ContainerStarted","Data":"e5d15d95d86459d776cd5a87de71836e59b4fc3f89340e54448996cdc214afc5"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.277174 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" podUID="e1488db8-3d22-4c1f-8400-8d292b48027a" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.288878 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" event={"ID":"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8","Type":"ContainerStarted","Data":"4fec17d7068e0d1d02a6f69274ff66f11e3f1b9b8ec71578543747402eb82801"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.300598 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6xfj\" (UniqueName: \"kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj\") pod \"redhat-operators-49m6v\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.303166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" event={"ID":"95e654ed-72c1-4f5b-b670-a89c96692455","Type":"ContainerStarted","Data":"f00a872b61bfb751f0f37220ff9661326b462d5d0af508d65c866d88a7ca8f58"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.304475 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" event={"ID":"9c72f60f-c311-45eb-84a1-8c95164be376","Type":"ContainerStarted","Data":"157cc07031add446052de24f811148255cd8b307febe8dcbeb00825b24a46f62"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.307406 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" event={"ID":"17c37ba2-ff55-4b60-a1db-01dec0522e2e","Type":"ContainerStarted","Data":"bc6deb6e63c78c86dde025958779bc4795db58ab7a976c2b4b1f48c6d4117254"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.307432 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" event={"ID":"17c37ba2-ff55-4b60-a1db-01dec0522e2e","Type":"ContainerStarted","Data":"5108ce05aea742f8b155eb2136f1b2400956040fa7d2426da2e96d24faedf61b"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.318585 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" podUID="17c37ba2-ff55-4b60-a1db-01dec0522e2e" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.335402 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerStarted","Data":"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b"} Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.335452 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerStarted","Data":"c2bc9942ad0793643bee9238e41da6673444abcb51aee18d6e32561967e7ae9c"} Oct 14 08:05:00 crc kubenswrapper[5002]: E1014 08:05:00.336610 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.364087 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:00 crc kubenswrapper[5002]: I1014 08:05:00.882773 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:00 crc kubenswrapper[5002]: W1014 08:05:00.896179 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c00ea43_0c55_40c3_8cf2_0f8727cc4e5a.slice/crio-f3e24d2cb162d998566efea0a7088163d12e1b43c7d00f65702499329b3975ec WatchSource:0}: Error finding container f3e24d2cb162d998566efea0a7088163d12e1b43c7d00f65702499329b3975ec: Status 404 returned error can't find the container with id f3e24d2cb162d998566efea0a7088163d12e1b43c7d00f65702499329b3975ec Oct 14 08:05:01 crc kubenswrapper[5002]: I1014 08:05:01.360348 5002 generic.go:334] "Generic (PLEG): container finished" podID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerID="4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a" exitCode=0 Oct 14 08:05:01 crc kubenswrapper[5002]: I1014 08:05:01.362304 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerDied","Data":"4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a"} Oct 14 08:05:01 crc kubenswrapper[5002]: I1014 08:05:01.362328 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerStarted","Data":"f3e24d2cb162d998566efea0a7088163d12e1b43c7d00f65702499329b3975ec"} Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.374448 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" podUID="42f32565-187e-4c71-a2c3-35eb988b8243" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.374768 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:ee05f2b06405240a8fcdbd430a9e8983b4667f372548334307b68c154e389960\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" podUID="e1488db8-3d22-4c1f-8400-8d292b48027a" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.374898 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" podUID="b04b0d3d-03a0-4f12-814f-599f2899962b" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.375004 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" podUID="b1921718-5347-4bfd-a327-9c38671b4888" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.375058 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" podUID="20fefb60-1453-4a27-9f7a-518df5b902fc" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.385019 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" Oct 14 08:05:01 crc kubenswrapper[5002]: E1014 08:05:01.401011 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" podUID="17c37ba2-ff55-4b60-a1db-01dec0522e2e" Oct 14 08:05:04 crc kubenswrapper[5002]: I1014 08:05:04.393972 5002 generic.go:334] "Generic (PLEG): container finished" podID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerID="bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc" exitCode=0 Oct 14 08:05:04 crc kubenswrapper[5002]: I1014 08:05:04.394083 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerDied","Data":"bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc"} Oct 14 08:05:08 crc kubenswrapper[5002]: I1014 08:05:08.828088 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-k6scp" Oct 14 08:05:15 crc kubenswrapper[5002]: E1014 08:05:15.590980 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867" Oct 14 08:05:15 crc kubenswrapper[5002]: E1014 08:05:15.592120 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m2kjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-7ffbcb7588-s6vkc_openstack-operators(bbc859e5-324a-413c-a7ac-499a0b0b8fba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:05:16 crc kubenswrapper[5002]: E1014 08:05:16.002386 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" podUID="bbc859e5-324a-413c-a7ac-499a0b0b8fba" Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.491860 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" event={"ID":"27e01872-38da-4f89-8232-c8edaec917f8","Type":"ContainerStarted","Data":"d99e110f6a00d556f5b9d6be90c6940c83614d49b1b0a259debe442344db4a61"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.496349 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" event={"ID":"cfe90f21-7b6a-4364-8c33-92b2880e9e0b","Type":"ContainerStarted","Data":"d0a17560f3f4b96e57d2bc2fffd4cb84021ff6fcab7f6dd5315b19d81a0cae56"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.498297 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" event={"ID":"05103074-be63-4240-bd7e-737c3c6ff4d5","Type":"ContainerStarted","Data":"260c6caf6a571974061c6ec5c6508e5dac1898e708f3e79a98ed0f6ce7ad67d8"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.513084 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" event={"ID":"95e654ed-72c1-4f5b-b670-a89c96692455","Type":"ContainerStarted","Data":"423156f449d10b64a85ce9dc281426f8fad25ee0b697aa5a9035b766b003276b"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.543981 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" event={"ID":"578becbc-f063-4b98-b5a3-eb0656b24850","Type":"ContainerStarted","Data":"cb90ecf86b67a262d6e8249ac6b3323f92ccf273bc446dcef95e7f5781732fda"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.547614 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" event={"ID":"8e0847dd-54d1-44f7-8df9-b1378b00cb67","Type":"ContainerStarted","Data":"2cf70f2a892f8ed6fa1581ca0bed5d00b1d1d1b76686933741eeaeb99d1af016"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.554350 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" event={"ID":"57a0f853-8feb-4064-9ed2-3460f2f66901","Type":"ContainerStarted","Data":"efc1cfd55a0b2cadb51f43585ef9e1d61f50298b68dd52a295549f5ebcd9299e"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.569777 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" event={"ID":"63c73302-dc00-4dfe-9d3f-d80cb181ffc2","Type":"ContainerStarted","Data":"b2f8d21afed5bde354ad8567cc0d1723579248105cfb0fdc75551db0a610a442"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.584476 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" event={"ID":"2041ec73-97de-43ea-ae4e-e8a1d7521157","Type":"ContainerStarted","Data":"d231c322026a113e855a56cd455826e96432fb4a3c3b25077aef09bf9516f8cd"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.590424 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerStarted","Data":"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5"} Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.605716 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" event={"ID":"bbc859e5-324a-413c-a7ac-499a0b0b8fba","Type":"ContainerStarted","Data":"e8da556458275a47305a6bb145ab9f6d9d8a4c1973c7afaedc91f2f4a6363f49"} Oct 14 08:05:16 crc kubenswrapper[5002]: E1014 08:05:16.607411 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" podUID="bbc859e5-324a-413c-a7ac-499a0b0b8fba" Oct 14 08:05:16 crc kubenswrapper[5002]: I1014 08:05:16.618640 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-49m6v" podStartSLOduration=2.262540722 podStartE2EDuration="16.618625615s" podCreationTimestamp="2025-10-14 08:05:00 +0000 UTC" firstStartedPulling="2025-10-14 08:05:01.374935802 +0000 UTC m=+834.356175264" lastFinishedPulling="2025-10-14 08:05:15.731020705 +0000 UTC m=+848.712260157" observedRunningTime="2025-10-14 08:05:16.616810177 +0000 UTC m=+849.598049639" watchObservedRunningTime="2025-10-14 08:05:16.618625615 +0000 UTC m=+849.599865067" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.614282 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" event={"ID":"2041ec73-97de-43ea-ae4e-e8a1d7521157","Type":"ContainerStarted","Data":"fafdf2a786ea01d28a1c8bd4ac1dd9037dfebe39e02498f23da1a382dd0ff133"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.614514 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.619795 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" event={"ID":"05103074-be63-4240-bd7e-737c3c6ff4d5","Type":"ContainerStarted","Data":"90cb0189fbbb8f5cbd3b767d1ee69a6ed8862bd58ad751d37042c3d4b6124097"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.619940 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.622546 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" event={"ID":"8e0847dd-54d1-44f7-8df9-b1378b00cb67","Type":"ContainerStarted","Data":"dcebfd98eed960a93318ca6d68b460a51dc6a84f6c3dd71915b61a94bac7f2f4"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.623259 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.627146 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" event={"ID":"63c73302-dc00-4dfe-9d3f-d80cb181ffc2","Type":"ContainerStarted","Data":"24f35674578cd76e25968bbe82127b71809f6cb5fb4717314b99cab3bee91fa8"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.627574 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.630084 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" event={"ID":"46630983-31be-46df-9e05-0506abbaf29e","Type":"ContainerStarted","Data":"a85bc67d1350f969f3d6d65143f598ebbc602ce32ea02eed1423841a1b98ccab"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.630110 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" event={"ID":"46630983-31be-46df-9e05-0506abbaf29e","Type":"ContainerStarted","Data":"190824be6cb48627585e4ee38489f08f5e6f751e07943db9a6b57b2e75977738"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.630448 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.632084 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" event={"ID":"de07a029-a377-47c8-be2f-a31148f4948a","Type":"ContainerStarted","Data":"434ad799d549ac293858e1d195ad0a46091947a5df8117c096e3beb884250cf0"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.632110 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" event={"ID":"de07a029-a377-47c8-be2f-a31148f4948a","Type":"ContainerStarted","Data":"cb0292ac026967b706d8879fe81eb8a8248d8cf10432d9e4cc85a5d67cc74006"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.632352 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.635916 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" podStartSLOduration=4.037081752 podStartE2EDuration="20.635904007s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.123851928 +0000 UTC m=+832.105091370" lastFinishedPulling="2025-10-14 08:05:15.722674143 +0000 UTC m=+848.703913625" observedRunningTime="2025-10-14 08:05:17.635101285 +0000 UTC m=+850.616340747" watchObservedRunningTime="2025-10-14 08:05:17.635904007 +0000 UTC m=+850.617143459" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.642005 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" event={"ID":"4b405ac3-bbcd-4432-8f99-2862ef406541","Type":"ContainerStarted","Data":"0a905fe8d750d80a4e33a8472df0f3007c76d59cae1afa2f306d31ab88fdbe35"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.642057 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" event={"ID":"4b405ac3-bbcd-4432-8f99-2862ef406541","Type":"ContainerStarted","Data":"7d6f6e615a417f75070a3ad737f5a475369bc842aadda13f958bc28943caa584"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.642132 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.650305 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" event={"ID":"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8","Type":"ContainerStarted","Data":"d5d9ba537658c5a9f7afe680096d6b5c3a41044a6517584add0252eb3119dd6d"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.650339 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" event={"ID":"74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8","Type":"ContainerStarted","Data":"4ca98a47068fabf089d6c0fa3881eb7e7ea013184cf4f69f2843742d7b89c332"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.650365 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.652565 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" podStartSLOduration=3.402519444 podStartE2EDuration="20.652555509s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.487159924 +0000 UTC m=+831.468399376" lastFinishedPulling="2025-10-14 08:05:15.737195989 +0000 UTC m=+848.718435441" observedRunningTime="2025-10-14 08:05:17.652024715 +0000 UTC m=+850.633264187" watchObservedRunningTime="2025-10-14 08:05:17.652555509 +0000 UTC m=+850.633794961" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.655532 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" event={"ID":"27e01872-38da-4f89-8232-c8edaec917f8","Type":"ContainerStarted","Data":"1b44470c7213324cc472ae2ce90a2a7176f3088438d8cf8465a45ff605054304"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.656049 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.665671 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" event={"ID":"cfe90f21-7b6a-4364-8c33-92b2880e9e0b","Type":"ContainerStarted","Data":"6cb8573d327ec3965a60679608df3191c6cdb5bd38f4baf7eaaf9a1e240dbab6"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.665825 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.686491 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" event={"ID":"9c72f60f-c311-45eb-84a1-8c95164be376","Type":"ContainerStarted","Data":"e4e93cdfbc5b2faa2b13191214e4bbf8a0b4a1acf351187d81fbfc48e56621cc"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.686769 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.686780 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" event={"ID":"9c72f60f-c311-45eb-84a1-8c95164be376","Type":"ContainerStarted","Data":"292ee4756fc7b8b090011cc49dabdb14b1372b2d950db7b837567727e76e660c"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.700967 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" podStartSLOduration=4.8344078889999995 podStartE2EDuration="20.700952547s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.855413126 +0000 UTC m=+832.836652568" lastFinishedPulling="2025-10-14 08:05:15.721957774 +0000 UTC m=+848.703197226" observedRunningTime="2025-10-14 08:05:17.686146493 +0000 UTC m=+850.667385965" watchObservedRunningTime="2025-10-14 08:05:17.700952547 +0000 UTC m=+850.682191999" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.706077 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" podStartSLOduration=3.486015907 podStartE2EDuration="20.706067164s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.486442846 +0000 UTC m=+831.467682298" lastFinishedPulling="2025-10-14 08:05:15.706494063 +0000 UTC m=+848.687733555" observedRunningTime="2025-10-14 08:05:17.703816394 +0000 UTC m=+850.685055856" watchObservedRunningTime="2025-10-14 08:05:17.706067164 +0000 UTC m=+850.687306616" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.710320 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" event={"ID":"578becbc-f063-4b98-b5a3-eb0656b24850","Type":"ContainerStarted","Data":"5d21ff371bc483f5840ef3363a151b02bc086bbf53e81bcd3936036fba3c7f3f"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.710936 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.716557 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" event={"ID":"95e654ed-72c1-4f5b-b670-a89c96692455","Type":"ContainerStarted","Data":"d6cdf950450ddb318e1a9b2db49fe172e75f5e79708849f0588a58510a8312ae"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.717147 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.725107 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" podStartSLOduration=3.740256472 podStartE2EDuration="20.72509142s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.747533303 +0000 UTC m=+831.728772755" lastFinishedPulling="2025-10-14 08:05:15.732368251 +0000 UTC m=+848.713607703" observedRunningTime="2025-10-14 08:05:17.721127494 +0000 UTC m=+850.702366956" watchObservedRunningTime="2025-10-14 08:05:17.72509142 +0000 UTC m=+850.706330872" Oct 14 08:05:17 crc kubenswrapper[5002]: E1014 08:05:17.732075 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:063a7e65b4ba98f0506f269ff7525b446eae06a5ed4a61c18ffa33a886500867\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" podUID="bbc859e5-324a-413c-a7ac-499a0b0b8fba" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.735815 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.735871 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" event={"ID":"57a0f853-8feb-4064-9ed2-3460f2f66901","Type":"ContainerStarted","Data":"73f9af24c27c16acc0e08a256ce6617e6a3764bacf5c8bb2a95639f230876692"} Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.742856 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" podStartSLOduration=3.771571316 podStartE2EDuration="20.742824212s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.736480249 +0000 UTC m=+831.717719701" lastFinishedPulling="2025-10-14 08:05:15.707733115 +0000 UTC m=+848.688972597" observedRunningTime="2025-10-14 08:05:17.738644841 +0000 UTC m=+850.719884303" watchObservedRunningTime="2025-10-14 08:05:17.742824212 +0000 UTC m=+850.724063664" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.784929 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" podStartSLOduration=4.211837852 podStartE2EDuration="20.784912992s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.163329878 +0000 UTC m=+832.144569330" lastFinishedPulling="2025-10-14 08:05:15.736404978 +0000 UTC m=+848.717644470" observedRunningTime="2025-10-14 08:05:17.783203737 +0000 UTC m=+850.764443189" watchObservedRunningTime="2025-10-14 08:05:17.784912992 +0000 UTC m=+850.766152444" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.785396 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" podStartSLOduration=3.82213982 podStartE2EDuration="20.785392314s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.743244319 +0000 UTC m=+831.724483771" lastFinishedPulling="2025-10-14 08:05:15.706496773 +0000 UTC m=+848.687736265" observedRunningTime="2025-10-14 08:05:17.760393469 +0000 UTC m=+850.741632921" watchObservedRunningTime="2025-10-14 08:05:17.785392314 +0000 UTC m=+850.766631766" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.828788 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" podStartSLOduration=4.431420784 podStartE2EDuration="20.828766888s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.344301314 +0000 UTC m=+832.325540766" lastFinishedPulling="2025-10-14 08:05:15.741647378 +0000 UTC m=+848.722886870" observedRunningTime="2025-10-14 08:05:17.827714681 +0000 UTC m=+850.808954133" watchObservedRunningTime="2025-10-14 08:05:17.828766888 +0000 UTC m=+850.810006340" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.832176 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" podStartSLOduration=3.599970238 podStartE2EDuration="20.832166479s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:58.48701643 +0000 UTC m=+831.468255882" lastFinishedPulling="2025-10-14 08:05:15.719212631 +0000 UTC m=+848.700452123" observedRunningTime="2025-10-14 08:05:17.806786814 +0000 UTC m=+850.788026276" watchObservedRunningTime="2025-10-14 08:05:17.832166479 +0000 UTC m=+850.813405931" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.871184 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" podStartSLOduration=4.349921296 podStartE2EDuration="20.871161197s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.215348212 +0000 UTC m=+832.196587664" lastFinishedPulling="2025-10-14 08:05:15.736588113 +0000 UTC m=+848.717827565" observedRunningTime="2025-10-14 08:05:17.866639696 +0000 UTC m=+850.847879158" watchObservedRunningTime="2025-10-14 08:05:17.871161197 +0000 UTC m=+850.852400649" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.886536 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" podStartSLOduration=4.38088161 podStartE2EDuration="20.886511846s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.214964042 +0000 UTC m=+832.196203494" lastFinishedPulling="2025-10-14 08:05:15.720594248 +0000 UTC m=+848.701833730" observedRunningTime="2025-10-14 08:05:17.882579431 +0000 UTC m=+850.863818903" watchObservedRunningTime="2025-10-14 08:05:17.886511846 +0000 UTC m=+850.867751298" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.901360 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" podStartSLOduration=4.408313481 podStartE2EDuration="20.90134572s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.215877117 +0000 UTC m=+832.197116559" lastFinishedPulling="2025-10-14 08:05:15.708909306 +0000 UTC m=+848.690148798" observedRunningTime="2025-10-14 08:05:17.89833566 +0000 UTC m=+850.879575112" watchObservedRunningTime="2025-10-14 08:05:17.90134572 +0000 UTC m=+850.882585172" Oct 14 08:05:17 crc kubenswrapper[5002]: I1014 08:05:17.922776 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" podStartSLOduration=4.3597703580000005 podStartE2EDuration="20.922757449s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.175183574 +0000 UTC m=+832.156423026" lastFinishedPulling="2025-10-14 08:05:15.738170635 +0000 UTC m=+848.719410117" observedRunningTime="2025-10-14 08:05:17.919333769 +0000 UTC m=+850.900573241" watchObservedRunningTime="2025-10-14 08:05:17.922757449 +0000 UTC m=+850.903996901" Oct 14 08:05:20 crc kubenswrapper[5002]: I1014 08:05:20.364882 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:20 crc kubenswrapper[5002]: I1014 08:05:20.365276 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:21 crc kubenswrapper[5002]: I1014 08:05:21.409459 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-49m6v" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="registry-server" probeResult="failure" output=< Oct 14 08:05:21 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 08:05:21 crc kubenswrapper[5002]: > Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.771648 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" event={"ID":"20fefb60-1453-4a27-9f7a-518df5b902fc","Type":"ContainerStarted","Data":"f9b40c16f9231c551ed3271669c5de40a7a8e25b3bc15379202fc6c0a5dc49b9"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.772291 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.774153 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" event={"ID":"b04b0d3d-03a0-4f12-814f-599f2899962b","Type":"ContainerStarted","Data":"dc8b80f939231c3294429cab8f2ef28cdbd2db2804814d5553abd52986bee288"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.774411 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.777737 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" event={"ID":"17c37ba2-ff55-4b60-a1db-01dec0522e2e","Type":"ContainerStarted","Data":"f4303ee1e66df1e8884224bc17c5f7422febe9c33252f26a19cfe42a753bf0a3"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.778167 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.780277 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerStarted","Data":"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.780517 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.782755 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" event={"ID":"e1488db8-3d22-4c1f-8400-8d292b48027a","Type":"ContainerStarted","Data":"1537eb843bb8420eb37e716902c9b322fa35f84c53a599d999f2311a22fd9519"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.783009 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.787367 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" event={"ID":"42f32565-187e-4c71-a2c3-35eb988b8243","Type":"ContainerStarted","Data":"aca9e4e89837b9e84eb73b61bbce56727862bfd8059ff7a56125003392d72f9c"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.787690 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.795251 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" event={"ID":"b1921718-5347-4bfd-a327-9c38671b4888","Type":"ContainerStarted","Data":"49591512877dab29bd3188e30f40227eca9a1ddaad70dfd801760739ff72baec"} Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.811047 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" podStartSLOduration=3.506488281 podStartE2EDuration="25.811030443s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.278991996 +0000 UTC m=+832.260231448" lastFinishedPulling="2025-10-14 08:05:21.583534148 +0000 UTC m=+854.564773610" observedRunningTime="2025-10-14 08:05:22.810483008 +0000 UTC m=+855.791722500" watchObservedRunningTime="2025-10-14 08:05:22.811030443 +0000 UTC m=+855.792269905" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.852668 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" podStartSLOduration=3.458913456 podStartE2EDuration="25.85264622s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.244987252 +0000 UTC m=+832.226226704" lastFinishedPulling="2025-10-14 08:05:21.638720006 +0000 UTC m=+854.619959468" observedRunningTime="2025-10-14 08:05:22.845350816 +0000 UTC m=+855.826590298" watchObservedRunningTime="2025-10-14 08:05:22.85264622 +0000 UTC m=+855.833885702" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.869521 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" podStartSLOduration=3.569005933 podStartE2EDuration="25.869496268s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.281262696 +0000 UTC m=+832.262502148" lastFinishedPulling="2025-10-14 08:05:21.581753021 +0000 UTC m=+854.562992483" observedRunningTime="2025-10-14 08:05:22.86841289 +0000 UTC m=+855.849652352" watchObservedRunningTime="2025-10-14 08:05:22.869496268 +0000 UTC m=+855.850735740" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.898944 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" podStartSLOduration=3.541902324 podStartE2EDuration="25.898928001s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.245175987 +0000 UTC m=+832.226415439" lastFinishedPulling="2025-10-14 08:05:21.602201654 +0000 UTC m=+854.583441116" observedRunningTime="2025-10-14 08:05:22.892902631 +0000 UTC m=+855.874142093" watchObservedRunningTime="2025-10-14 08:05:22.898928001 +0000 UTC m=+855.880167463" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.916041 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" podStartSLOduration=3.551017086 podStartE2EDuration="25.916017037s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.216722789 +0000 UTC m=+832.197962231" lastFinishedPulling="2025-10-14 08:05:21.58172274 +0000 UTC m=+854.562962182" observedRunningTime="2025-10-14 08:05:22.913380697 +0000 UTC m=+855.894620169" watchObservedRunningTime="2025-10-14 08:05:22.916017037 +0000 UTC m=+855.897256529" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.957063 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podStartSLOduration=3.622526059 podStartE2EDuration="25.957040108s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.249041779 +0000 UTC m=+832.230281231" lastFinishedPulling="2025-10-14 08:05:21.583555788 +0000 UTC m=+854.564795280" observedRunningTime="2025-10-14 08:05:22.930007259 +0000 UTC m=+855.911246741" watchObservedRunningTime="2025-10-14 08:05:22.957040108 +0000 UTC m=+855.938279570" Oct 14 08:05:22 crc kubenswrapper[5002]: I1014 08:05:22.960565 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-s89dt" podStartSLOduration=3.658764943 podStartE2EDuration="25.960549111s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.28174861 +0000 UTC m=+832.262988052" lastFinishedPulling="2025-10-14 08:05:21.583532728 +0000 UTC m=+854.564772220" observedRunningTime="2025-10-14 08:05:22.950293679 +0000 UTC m=+855.931533131" watchObservedRunningTime="2025-10-14 08:05:22.960549111 +0000 UTC m=+855.941788563" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.406944 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-qpnq7" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.433185 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-9kfv7" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.455008 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lr6gb" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.526300 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-8j2qn" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.610747 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-5t4hm" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.625117 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-69rk7" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.651623 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-xsb2d" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.753936 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-82c6h" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.819331 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-cn7pq" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.840952 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-xj9vc" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.846090 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-7jcbw" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.896924 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-59jzf" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.901693 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-kqdfb" Oct 14 08:05:27 crc kubenswrapper[5002]: I1014 08:05:27.989254 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-74c8t" Oct 14 08:05:28 crc kubenswrapper[5002]: I1014 08:05:28.052545 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-fp4rw" Oct 14 08:05:28 crc kubenswrapper[5002]: I1014 08:05:28.092733 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:05:28 crc kubenswrapper[5002]: I1014 08:05:28.157667 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-d9ftz" Oct 14 08:05:28 crc kubenswrapper[5002]: I1014 08:05:28.171306 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-lwlwj" Oct 14 08:05:28 crc kubenswrapper[5002]: I1014 08:05:28.186041 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-tws79" Oct 14 08:05:29 crc kubenswrapper[5002]: I1014 08:05:29.357084 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d44848n4xg8" Oct 14 08:05:30 crc kubenswrapper[5002]: I1014 08:05:30.433658 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:30 crc kubenswrapper[5002]: I1014 08:05:30.493794 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:30 crc kubenswrapper[5002]: I1014 08:05:30.687439 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:31 crc kubenswrapper[5002]: I1014 08:05:31.724564 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:05:31 crc kubenswrapper[5002]: I1014 08:05:31.873372 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-49m6v" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="registry-server" containerID="cri-o://ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5" gracePeriod=2 Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.353480 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.414208 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content\") pod \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.515570 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities\") pod \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.515665 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6xfj\" (UniqueName: \"kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj\") pod \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\" (UID: \"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a\") " Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.518221 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities" (OuterVolumeSpecName: "utilities") pod "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" (UID: "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.532770 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj" (OuterVolumeSpecName: "kube-api-access-f6xfj") pod "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" (UID: "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a"). InnerVolumeSpecName "kube-api-access-f6xfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.540960 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" (UID: "7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.617230 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.617678 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6xfj\" (UniqueName: \"kubernetes.io/projected/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-kube-api-access-f6xfj\") on node \"crc\" DevicePath \"\"" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.617700 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.884253 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" event={"ID":"bbc859e5-324a-413c-a7ac-499a0b0b8fba","Type":"ContainerStarted","Data":"cc6e40e25786bb4fcbfd66dc4d840aaca41e659103b62211beb591bd50ad1b40"} Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.886116 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.888981 5002 generic.go:334] "Generic (PLEG): container finished" podID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerID="ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5" exitCode=0 Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.889034 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerDied","Data":"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5"} Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.889065 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49m6v" event={"ID":"7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a","Type":"ContainerDied","Data":"f3e24d2cb162d998566efea0a7088163d12e1b43c7d00f65702499329b3975ec"} Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.889093 5002 scope.go:117] "RemoveContainer" containerID="ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.889273 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49m6v" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.921981 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" podStartSLOduration=2.804241453 podStartE2EDuration="35.921953019s" podCreationTimestamp="2025-10-14 08:04:57 +0000 UTC" firstStartedPulling="2025-10-14 08:04:59.148302778 +0000 UTC m=+832.129542230" lastFinishedPulling="2025-10-14 08:05:32.266014344 +0000 UTC m=+865.247253796" observedRunningTime="2025-10-14 08:05:32.915910129 +0000 UTC m=+865.897149641" watchObservedRunningTime="2025-10-14 08:05:32.921953019 +0000 UTC m=+865.903192501" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.934960 5002 scope.go:117] "RemoveContainer" containerID="bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.961310 5002 scope.go:117] "RemoveContainer" containerID="4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.963926 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.969370 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-49m6v"] Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.987952 5002 scope.go:117] "RemoveContainer" containerID="ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5" Oct 14 08:05:32 crc kubenswrapper[5002]: E1014 08:05:32.999263 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5\": container with ID starting with ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5 not found: ID does not exist" containerID="ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.999308 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5"} err="failed to get container status \"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5\": rpc error: code = NotFound desc = could not find container \"ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5\": container with ID starting with ba7ab0c097889b16277b84b1ebf1d25e8296c4d826df426d9d1d8f5db9dbc3e5 not found: ID does not exist" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.999337 5002 scope.go:117] "RemoveContainer" containerID="bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc" Oct 14 08:05:32 crc kubenswrapper[5002]: E1014 08:05:32.999624 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc\": container with ID starting with bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc not found: ID does not exist" containerID="bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.999678 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc"} err="failed to get container status \"bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc\": rpc error: code = NotFound desc = could not find container \"bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc\": container with ID starting with bd8d8cf51a383def00193f60ab2050d15e0ddadd6699c6ba760588204bfa2ddc not found: ID does not exist" Oct 14 08:05:32 crc kubenswrapper[5002]: I1014 08:05:32.999712 5002 scope.go:117] "RemoveContainer" containerID="4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a" Oct 14 08:05:33 crc kubenswrapper[5002]: E1014 08:05:33.000091 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a\": container with ID starting with 4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a not found: ID does not exist" containerID="4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a" Oct 14 08:05:33 crc kubenswrapper[5002]: I1014 08:05:33.000115 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a"} err="failed to get container status \"4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a\": rpc error: code = NotFound desc = could not find container \"4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a\": container with ID starting with 4db3d75ac2c971e649593b7545e3c68f1451b398f5ad42d128e0ba869320337a not found: ID does not exist" Oct 14 08:05:33 crc kubenswrapper[5002]: I1014 08:05:33.736040 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" path="/var/lib/kubelet/pods/7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a/volumes" Oct 14 08:05:37 crc kubenswrapper[5002]: I1014 08:05:37.680320 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-s6vkc" Oct 14 08:05:39 crc kubenswrapper[5002]: I1014 08:05:39.218406 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:05:39 crc kubenswrapper[5002]: I1014 08:05:39.218544 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:05:54 crc kubenswrapper[5002]: I1014 08:05:54.819916 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:05:55 crc kubenswrapper[5002]: E1014 08:05:54.831011 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="extract-content" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.831058 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="extract-content" Oct 14 08:05:55 crc kubenswrapper[5002]: E1014 08:05:54.831121 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="registry-server" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.831133 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="registry-server" Oct 14 08:05:55 crc kubenswrapper[5002]: E1014 08:05:54.831148 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="extract-utilities" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.831158 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="extract-utilities" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.832107 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c00ea43-0c55-40c3-8cf2-0f8727cc4e5a" containerName="registry-server" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.833353 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.836159 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-zrz2q" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.837260 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.837387 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.837506 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.841678 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.857753 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcfq7\" (UniqueName: \"kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.857889 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.899082 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.905171 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.909212 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.915504 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.958636 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.958703 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.959297 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96zcx\" (UniqueName: \"kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.959355 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcfq7\" (UniqueName: \"kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.959505 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.960047 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:54.979397 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcfq7\" (UniqueName: \"kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7\") pod \"dnsmasq-dns-5d487d97d7-6v52n\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.060807 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.060875 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.060918 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96zcx\" (UniqueName: \"kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.063366 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.067191 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.082589 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96zcx\" (UniqueName: \"kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx\") pod \"dnsmasq-dns-6948694bd9-5xrxd\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.158269 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.238758 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.554424 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:05:55 crc kubenswrapper[5002]: I1014 08:05:55.710951 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:05:56 crc kubenswrapper[5002]: I1014 08:05:56.085993 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" event={"ID":"1764ec79-d795-4857-a364-1f2326a0291e","Type":"ContainerStarted","Data":"3f594c8a3adf8fab4a9731d8a0627b8a25b266533203fb19d03d45eefb750ccf"} Oct 14 08:05:56 crc kubenswrapper[5002]: I1014 08:05:56.087723 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" event={"ID":"d7fae819-1eef-4b7e-a933-34ba26155759","Type":"ContainerStarted","Data":"52f2a3c36a7cd3068adc5ae3a7df981c5c40247fa605d27c14d2fa41c2db268d"} Oct 14 08:05:57 crc kubenswrapper[5002]: I1014 08:05:57.884341 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:05:57 crc kubenswrapper[5002]: I1014 08:05:57.908293 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:05:57 crc kubenswrapper[5002]: I1014 08:05:57.909469 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:57 crc kubenswrapper[5002]: I1014 08:05:57.915255 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.105996 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.106143 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.106178 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tww96\" (UniqueName: \"kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.126429 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.161071 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.162712 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.174231 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.207544 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.207657 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tww96\" (UniqueName: \"kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.207679 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.208763 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.208820 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.253622 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tww96\" (UniqueName: \"kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96\") pod \"dnsmasq-dns-586b486d97-dcpvh\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.311965 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tm44\" (UniqueName: \"kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.312273 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.312293 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.413478 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tm44\" (UniqueName: \"kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.413601 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.414465 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.414594 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.416297 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.438675 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tm44\" (UniqueName: \"kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44\") pod \"dnsmasq-dns-7869c47d6c-55tk2\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.490142 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.532311 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:05:58 crc kubenswrapper[5002]: I1014 08:05:58.963856 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.037379 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.041058 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.043568 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.049019 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.049133 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-z9926" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.049248 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.049501 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.049845 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.050619 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.050674 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.067741 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.224910 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.224970 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225016 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hmj4\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225220 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225305 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225343 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225391 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225535 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225566 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225588 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.225603 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.288118 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.289716 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.295592 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.296059 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.296227 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tgsb8" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.296368 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.296482 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.297115 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.297286 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.302485 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.326878 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hmj4\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.326939 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.326963 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.326981 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327025 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327047 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327070 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327116 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327132 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327180 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.327219 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.328006 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.328669 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.329663 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.329984 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.330081 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.332417 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.334575 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.334764 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.335377 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.337373 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.344479 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hmj4\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.356361 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.363941 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429544 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbsz8\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429597 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429645 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429671 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429688 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429704 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429769 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429789 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429812 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.429830 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.431071 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.533383 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbsz8\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.534965 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535413 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535563 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535589 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535602 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535635 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535658 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535677 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535701 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535741 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.535942 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.536261 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.536274 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.537134 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.539020 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.539273 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.542694 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.542988 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.544205 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.546781 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.553132 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbsz8\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.558941 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " pod="openstack/rabbitmq-server-0" Oct 14 08:05:59 crc kubenswrapper[5002]: I1014 08:05:59.622398 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.745519 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.747649 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.749909 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.750751 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.752072 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8csx9" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.752257 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.756405 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.761146 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.765329 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852524 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852573 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-config-data-default\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852660 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852688 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/355bb6ac-9bd2-411e-81d4-400069911641-config-data-generated\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852706 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x2l7\" (UniqueName: \"kubernetes.io/projected/355bb6ac-9bd2-411e-81d4-400069911641-kube-api-access-8x2l7\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852777 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-secrets\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852817 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-operator-scripts\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.852871 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-kolla-config\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.955611 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-secrets\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.955719 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-operator-scripts\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.955832 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-kolla-config\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.955973 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.956028 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-config-data-default\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.956091 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.956215 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.956287 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/355bb6ac-9bd2-411e-81d4-400069911641-config-data-generated\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.956332 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x2l7\" (UniqueName: \"kubernetes.io/projected/355bb6ac-9bd2-411e-81d4-400069911641-kube-api-access-8x2l7\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.958413 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/355bb6ac-9bd2-411e-81d4-400069911641-config-data-generated\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.959551 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.959925 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-kolla-config\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.960300 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-config-data-default\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.961374 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/355bb6ac-9bd2-411e-81d4-400069911641-operator-scripts\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:00 crc kubenswrapper[5002]: I1014 08:06:00.969659 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-secrets\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:01 crc kubenswrapper[5002]: I1014 08:06:01.008489 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:01 crc kubenswrapper[5002]: I1014 08:06:01.013273 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/355bb6ac-9bd2-411e-81d4-400069911641-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:01 crc kubenswrapper[5002]: I1014 08:06:01.021053 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x2l7\" (UniqueName: \"kubernetes.io/projected/355bb6ac-9bd2-411e-81d4-400069911641-kube-api-access-8x2l7\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:01 crc kubenswrapper[5002]: I1014 08:06:01.022229 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"355bb6ac-9bd2-411e-81d4-400069911641\") " pod="openstack/openstack-galera-0" Oct 14 08:06:01 crc kubenswrapper[5002]: I1014 08:06:01.086080 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.181738 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" event={"ID":"190e7b2c-275b-4b93-8485-46659be96268","Type":"ContainerStarted","Data":"686dd9c40a57d91f290877e05593947acfd293ae831cc59fef7dc07055ee7006"} Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.238724 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.243058 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.250778 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.288249 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.288338 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.288247 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7j2l5" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.288688 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.386284 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66jhr\" (UniqueName: \"kubernetes.io/projected/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kube-api-access-66jhr\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.388996 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.389229 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.389259 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.390054 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.390114 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.390140 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.390310 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.390661 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.491883 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.491946 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.491974 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66jhr\" (UniqueName: \"kubernetes.io/projected/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kube-api-access-66jhr\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.491999 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.492030 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.492048 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.492089 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.492118 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.492139 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.493577 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.493977 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.494119 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.494285 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.495389 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.501031 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.503103 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.503699 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.527608 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66jhr\" (UniqueName: \"kubernetes.io/projected/2c1f1d79-cecd-4eda-a876-11dfb83ab76b-kube-api-access-66jhr\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.531266 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2c1f1d79-cecd-4eda-a876-11dfb83ab76b\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.591310 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.592491 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.594427 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.594657 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.594778 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2skm2" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.601232 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.613029 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.694186 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w2vk\" (UniqueName: \"kubernetes.io/projected/d0a98c43-f790-476c-822c-e9b5957bbb78-kube-api-access-2w2vk\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.694327 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.694546 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-config-data\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.694620 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.694653 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-kolla-config\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.795911 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w2vk\" (UniqueName: \"kubernetes.io/projected/d0a98c43-f790-476c-822c-e9b5957bbb78-kube-api-access-2w2vk\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.795982 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.796048 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-config-data\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.796066 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.796080 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-kolla-config\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.798388 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-kolla-config\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.798443 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0a98c43-f790-476c-822c-e9b5957bbb78-config-data\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.800290 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.800451 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0a98c43-f790-476c-822c-e9b5957bbb78-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.819805 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w2vk\" (UniqueName: \"kubernetes.io/projected/d0a98c43-f790-476c-822c-e9b5957bbb78-kube-api-access-2w2vk\") pod \"memcached-0\" (UID: \"d0a98c43-f790-476c-822c-e9b5957bbb78\") " pod="openstack/memcached-0" Oct 14 08:06:02 crc kubenswrapper[5002]: I1014 08:06:02.925300 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:06:03 crc kubenswrapper[5002]: I1014 08:06:03.192323 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" event={"ID":"3699a03e-0a96-40d8-ac3c-effd67d9729c","Type":"ContainerStarted","Data":"c32ce2244c8ac7f3621401133b86818e0c27f92c6d06b0adfe831f91648195bd"} Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.671374 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.672588 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.675081 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vq5jr" Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.685028 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.839040 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5vrq\" (UniqueName: \"kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq\") pod \"kube-state-metrics-0\" (UID: \"3fd0a0da-37d6-424c-b8ad-87791d2b155e\") " pod="openstack/kube-state-metrics-0" Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.940478 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5vrq\" (UniqueName: \"kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq\") pod \"kube-state-metrics-0\" (UID: \"3fd0a0da-37d6-424c-b8ad-87791d2b155e\") " pod="openstack/kube-state-metrics-0" Oct 14 08:06:04 crc kubenswrapper[5002]: I1014 08:06:04.964801 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5vrq\" (UniqueName: \"kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq\") pod \"kube-state-metrics-0\" (UID: \"3fd0a0da-37d6-424c-b8ad-87791d2b155e\") " pod="openstack/kube-state-metrics-0" Oct 14 08:06:05 crc kubenswrapper[5002]: I1014 08:06:05.018322 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.887673 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tw5d7"] Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.899004 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.911579 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7"] Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.918293 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.918617 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.919699 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-n5j7z"] Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.921150 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dpbw7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.921710 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.931655 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-n5j7z"] Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.996886 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-log-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.996990 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-combined-ca-bundle\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.997024 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzg6s\" (UniqueName: \"kubernetes.io/projected/f9439452-e42f-4fa8-ac61-65b5389ff828-kube-api-access-kzg6s\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.997127 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.997156 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-ovn-controller-tls-certs\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.997239 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9439452-e42f-4fa8-ac61-65b5389ff828-scripts\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:07 crc kubenswrapper[5002]: I1014 08:06:07.997275 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098395 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-run\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098464 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-combined-ca-bundle\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098495 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-lib\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098518 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzg6s\" (UniqueName: \"kubernetes.io/projected/f9439452-e42f-4fa8-ac61-65b5389ff828-kube-api-access-kzg6s\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098534 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82g4m\" (UniqueName: \"kubernetes.io/projected/4074edb0-c770-4819-a4bd-581f3e7c6e23-kube-api-access-82g4m\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098585 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098605 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-ovn-controller-tls-certs\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098639 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9439452-e42f-4fa8-ac61-65b5389ff828-scripts\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098666 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098698 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4074edb0-c770-4819-a4bd-581f3e7c6e23-scripts\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098724 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-log-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098757 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-etc-ovs\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.098782 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-log\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.099617 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.099713 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-run\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.099755 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f9439452-e42f-4fa8-ac61-65b5389ff828-var-log-ovn\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.101319 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f9439452-e42f-4fa8-ac61-65b5389ff828-scripts\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.103157 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-combined-ca-bundle\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.103235 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9439452-e42f-4fa8-ac61-65b5389ff828-ovn-controller-tls-certs\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.117638 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzg6s\" (UniqueName: \"kubernetes.io/projected/f9439452-e42f-4fa8-ac61-65b5389ff828-kube-api-access-kzg6s\") pod \"ovn-controller-tw5d7\" (UID: \"f9439452-e42f-4fa8-ac61-65b5389ff828\") " pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199653 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4074edb0-c770-4819-a4bd-581f3e7c6e23-scripts\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199697 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-etc-ovs\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199715 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-log\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199736 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-run\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199774 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-lib\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.199792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82g4m\" (UniqueName: \"kubernetes.io/projected/4074edb0-c770-4819-a4bd-581f3e7c6e23-kube-api-access-82g4m\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.200213 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-run\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.204709 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-etc-ovs\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.206455 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-log\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.206531 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4074edb0-c770-4819-a4bd-581f3e7c6e23-var-lib\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.208604 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4074edb0-c770-4819-a4bd-581f3e7c6e23-scripts\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.225257 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82g4m\" (UniqueName: \"kubernetes.io/projected/4074edb0-c770-4819-a4bd-581f3e7c6e23-kube-api-access-82g4m\") pod \"ovn-controller-ovs-n5j7z\" (UID: \"4074edb0-c770-4819-a4bd-581f3e7c6e23\") " pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.248203 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:08 crc kubenswrapper[5002]: I1014 08:06:08.258235 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:09 crc kubenswrapper[5002]: I1014 08:06:09.218823 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:06:09 crc kubenswrapper[5002]: I1014 08:06:09.218923 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.947869 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.948298 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jcfq7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d487d97d7-6v52n_openstack(1764ec79-d795-4857-a364-1f2326a0291e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.949453 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" podUID="1764ec79-d795-4857-a364-1f2326a0291e" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.953540 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.953791 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-96zcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6948694bd9-5xrxd_openstack(d7fae819-1eef-4b7e-a933-34ba26155759): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:06:10 crc kubenswrapper[5002]: E1014 08:06:10.954996 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" podUID="d7fae819-1eef-4b7e-a933-34ba26155759" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.449695 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.474447 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: W1014 08:06:11.490605 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb16df05d_7859_4d03_8a4b_ca31e68d1d32.slice/crio-44d459f7ed92756cbebddc17016393b9d510953fefcd538061bff7d3d4269d07 WatchSource:0}: Error finding container 44d459f7ed92756cbebddc17016393b9d510953fefcd538061bff7d3d4269d07: Status 404 returned error can't find the container with id 44d459f7ed92756cbebddc17016393b9d510953fefcd538061bff7d3d4269d07 Oct 14 08:06:11 crc kubenswrapper[5002]: W1014 08:06:11.630362 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7fe047a_78a9_4abf_a948_167ba509729c.slice/crio-0cb81a0212979b5d88f7ff7c929cd5f4ae717239fdb60a73e486b89f050ab982 WatchSource:0}: Error finding container 0cb81a0212979b5d88f7ff7c929cd5f4ae717239fdb60a73e486b89f050ab982: Status 404 returned error can't find the container with id 0cb81a0212979b5d88f7ff7c929cd5f4ae717239fdb60a73e486b89f050ab982 Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.647268 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.778947 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7"] Oct 14 08:06:11 crc kubenswrapper[5002]: W1014 08:06:11.783002 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fd0a0da_37d6_424c_b8ad_87791d2b155e.slice/crio-c41ec304711e83e77f3aeab30b2209511ac236c34d4d76d7a17b44fba0a60ede WatchSource:0}: Error finding container c41ec304711e83e77f3aeab30b2209511ac236c34d4d76d7a17b44fba0a60ede: Status 404 returned error can't find the container with id c41ec304711e83e77f3aeab30b2209511ac236c34d4d76d7a17b44fba0a60ede Oct 14 08:06:11 crc kubenswrapper[5002]: W1014 08:06:11.783706 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9439452_e42f_4fa8_ac61_65b5389ff828.slice/crio-e9d1a817613b4c4a50e200189d9a8ee44924f871192aa98d7c39f2c393a3ed61 WatchSource:0}: Error finding container e9d1a817613b4c4a50e200189d9a8ee44924f871192aa98d7c39f2c393a3ed61: Status 404 returned error can't find the container with id e9d1a817613b4c4a50e200189d9a8ee44924f871192aa98d7c39f2c393a3ed61 Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.791123 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.808876 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.812045 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.815141 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.832199 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.879614 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-n5j7z"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.910725 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.916720 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.921565 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-lkr5z" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.921742 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.921865 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.922042 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.922671 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.925864 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.972658 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96zcx\" (UniqueName: \"kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx\") pod \"d7fae819-1eef-4b7e-a933-34ba26155759\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.972731 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcfq7\" (UniqueName: \"kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7\") pod \"1764ec79-d795-4857-a364-1f2326a0291e\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.972811 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config\") pod \"1764ec79-d795-4857-a364-1f2326a0291e\" (UID: \"1764ec79-d795-4857-a364-1f2326a0291e\") " Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.973457 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config" (OuterVolumeSpecName: "config") pod "1764ec79-d795-4857-a364-1f2326a0291e" (UID: "1764ec79-d795-4857-a364-1f2326a0291e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.973505 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config\") pod \"d7fae819-1eef-4b7e-a933-34ba26155759\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.973748 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config" (OuterVolumeSpecName: "config") pod "d7fae819-1eef-4b7e-a933-34ba26155759" (UID: "d7fae819-1eef-4b7e-a933-34ba26155759"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.973868 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc\") pod \"d7fae819-1eef-4b7e-a933-34ba26155759\" (UID: \"d7fae819-1eef-4b7e-a933-34ba26155759\") " Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.974180 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1764ec79-d795-4857-a364-1f2326a0291e-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.974198 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.974541 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7fae819-1eef-4b7e-a933-34ba26155759" (UID: "d7fae819-1eef-4b7e-a933-34ba26155759"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.978235 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx" (OuterVolumeSpecName: "kube-api-access-96zcx") pod "d7fae819-1eef-4b7e-a933-34ba26155759" (UID: "d7fae819-1eef-4b7e-a933-34ba26155759"). InnerVolumeSpecName "kube-api-access-96zcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:11 crc kubenswrapper[5002]: I1014 08:06:11.978776 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7" (OuterVolumeSpecName: "kube-api-access-jcfq7") pod "1764ec79-d795-4857-a364-1f2326a0291e" (UID: "1764ec79-d795-4857-a364-1f2326a0291e"). InnerVolumeSpecName "kube-api-access-jcfq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076394 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076488 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076552 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076641 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-config\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076879 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4qpx\" (UniqueName: \"kubernetes.io/projected/705e9cf2-0759-49ed-8407-a73c3aed75cf-kube-api-access-m4qpx\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.076933 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.077049 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.077300 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96zcx\" (UniqueName: \"kubernetes.io/projected/d7fae819-1eef-4b7e-a933-34ba26155759-kube-api-access-96zcx\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.077329 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcfq7\" (UniqueName: \"kubernetes.io/projected/1764ec79-d795-4857-a364-1f2326a0291e-kube-api-access-jcfq7\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.077339 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7fae819-1eef-4b7e-a933-34ba26155759-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.097250 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.099647 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.101763 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.102303 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.102564 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.104444 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-68fw2" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.124314 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.178780 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-config\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.178948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179045 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179096 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179145 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179178 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179218 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179254 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-config\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179307 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179354 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179421 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4qpx\" (UniqueName: \"kubernetes.io/projected/705e9cf2-0759-49ed-8407-a73c3aed75cf-kube-api-access-m4qpx\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179453 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179501 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179533 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179572 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.179604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgxpg\" (UniqueName: \"kubernetes.io/projected/2b959df9-bb11-4ee8-8af1-73beac51cfba-kube-api-access-dgxpg\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.180093 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.181397 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-config\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.182937 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/705e9cf2-0759-49ed-8407-a73c3aed75cf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.183957 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.185431 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.185974 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.188376 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/705e9cf2-0759-49ed-8407-a73c3aed75cf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.211460 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4qpx\" (UniqueName: \"kubernetes.io/projected/705e9cf2-0759-49ed-8407-a73c3aed75cf-kube-api-access-m4qpx\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.213405 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"705e9cf2-0759-49ed-8407-a73c3aed75cf\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.244742 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281626 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281693 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281745 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgxpg\" (UniqueName: \"kubernetes.io/projected/2b959df9-bb11-4ee8-8af1-73beac51cfba-kube-api-access-dgxpg\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281793 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-config\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281867 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.281988 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.282058 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.282099 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.283708 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.284441 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.285691 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.286000 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b959df9-bb11-4ee8-8af1-73beac51cfba-config\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.286772 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-n5j7z" event={"ID":"4074edb0-c770-4819-a4bd-581f3e7c6e23","Type":"ContainerStarted","Data":"54511cf7d4ecdc00fbb901a5dd0fa6a7b8741cf9e6191603512d2a5cc3469182"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.289249 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerStarted","Data":"44d459f7ed92756cbebddc17016393b9d510953fefcd538061bff7d3d4269d07"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.289806 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.289811 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.291068 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b959df9-bb11-4ee8-8af1-73beac51cfba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.292322 5002 generic.go:334] "Generic (PLEG): container finished" podID="190e7b2c-275b-4b93-8485-46659be96268" containerID="3120c99ba207b350798c968c9a2639ebeb0dc1a22a36374d90a8273f9c3e13be" exitCode=0 Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.292544 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" event={"ID":"190e7b2c-275b-4b93-8485-46659be96268","Type":"ContainerDied","Data":"3120c99ba207b350798c968c9a2639ebeb0dc1a22a36374d90a8273f9c3e13be"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.294467 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2c1f1d79-cecd-4eda-a876-11dfb83ab76b","Type":"ContainerStarted","Data":"f518c1c38d086f476fcf91c2dcab3dceedf3796ea424ec930f895b4dcd27b6b9"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.295633 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" event={"ID":"1764ec79-d795-4857-a364-1f2326a0291e","Type":"ContainerDied","Data":"3f594c8a3adf8fab4a9731d8a0627b8a25b266533203fb19d03d45eefb750ccf"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.295692 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-6v52n" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.301792 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"355bb6ac-9bd2-411e-81d4-400069911641","Type":"ContainerStarted","Data":"894c67e36814417c9c5e9bec1661740dc21f6f3f0a020ca91e4afbf4a72731a7"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.303240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fd0a0da-37d6-424c-b8ad-87791d2b155e","Type":"ContainerStarted","Data":"c41ec304711e83e77f3aeab30b2209511ac236c34d4d76d7a17b44fba0a60ede"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.307311 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d0a98c43-f790-476c-822c-e9b5957bbb78","Type":"ContainerStarted","Data":"b7586e52b115c9dc000018daa356fa673bb3a107c0ef38e7d94667b0959ee791"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.314050 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgxpg\" (UniqueName: \"kubernetes.io/projected/2b959df9-bb11-4ee8-8af1-73beac51cfba-kube-api-access-dgxpg\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.319275 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" event={"ID":"d7fae819-1eef-4b7e-a933-34ba26155759","Type":"ContainerDied","Data":"52f2a3c36a7cd3068adc5ae3a7df981c5c40247fa605d27c14d2fa41c2db268d"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.319315 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5xrxd" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.325393 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerStarted","Data":"0cb81a0212979b5d88f7ff7c929cd5f4ae717239fdb60a73e486b89f050ab982"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.329552 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7" event={"ID":"f9439452-e42f-4fa8-ac61-65b5389ff828","Type":"ContainerStarted","Data":"e9d1a817613b4c4a50e200189d9a8ee44924f871192aa98d7c39f2c393a3ed61"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.334550 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"2b959df9-bb11-4ee8-8af1-73beac51cfba\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.334568 5002 generic.go:334] "Generic (PLEG): container finished" podID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerID="5e9d56ec78f3e463a8b0e9a06bb0372a93e6ddf842e85e4b4067c6ac574fd867" exitCode=0 Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.334611 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" event={"ID":"3699a03e-0a96-40d8-ac3c-effd67d9729c","Type":"ContainerDied","Data":"5e9d56ec78f3e463a8b0e9a06bb0372a93e6ddf842e85e4b4067c6ac574fd867"} Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.370065 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.377714 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-6v52n"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.391384 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.397803 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5xrxd"] Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.429056 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:12 crc kubenswrapper[5002]: I1014 08:06:12.823655 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.027795 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-b85t2"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.035339 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.039581 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.073612 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-b85t2"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.166134 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.203609 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.217743 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218400 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-combined-ca-bundle\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218439 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218497 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhnn8\" (UniqueName: \"kubernetes.io/projected/8622c6ea-480a-457b-b42c-2adad7acc425-kube-api-access-zhnn8\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218542 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovs-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218573 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovn-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.218602 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8622c6ea-480a-457b-b42c-2adad7acc425-config\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.219090 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.222985 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.224101 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320173 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-combined-ca-bundle\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320217 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320279 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhnn8\" (UniqueName: \"kubernetes.io/projected/8622c6ea-480a-457b-b42c-2adad7acc425-kube-api-access-zhnn8\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320307 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5627s\" (UniqueName: \"kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320330 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320368 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320385 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovs-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320409 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320428 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovn-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.320462 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8622c6ea-480a-457b-b42c-2adad7acc425-config\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.321235 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8622c6ea-480a-457b-b42c-2adad7acc425-config\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.321418 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovs-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.321481 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8622c6ea-480a-457b-b42c-2adad7acc425-ovn-rundir\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.327083 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-combined-ca-bundle\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.330618 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8622c6ea-480a-457b-b42c-2adad7acc425-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.341365 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhnn8\" (UniqueName: \"kubernetes.io/projected/8622c6ea-480a-457b-b42c-2adad7acc425-kube-api-access-zhnn8\") pod \"ovn-controller-metrics-b85t2\" (UID: \"8622c6ea-480a-457b-b42c-2adad7acc425\") " pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.366381 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" event={"ID":"190e7b2c-275b-4b93-8485-46659be96268","Type":"ContainerStarted","Data":"a482acc915c7fdd1cb544a60b8d910573b0396ead0a4ee90e0e85e1de922c709"} Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.366456 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.368510 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"705e9cf2-0759-49ed-8407-a73c3aed75cf","Type":"ContainerStarted","Data":"248b5b9d426364db6b13665bb12a78a7221dc0572f44c53e94da59bd6aec88e1"} Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.382038 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-b85t2" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.382274 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" podStartSLOduration=5.934179574 podStartE2EDuration="15.382255799s" podCreationTimestamp="2025-10-14 08:05:58 +0000 UTC" firstStartedPulling="2025-10-14 08:06:01.667257338 +0000 UTC m=+894.648496800" lastFinishedPulling="2025-10-14 08:06:11.115333573 +0000 UTC m=+904.096573025" observedRunningTime="2025-10-14 08:06:13.382036303 +0000 UTC m=+906.363275755" watchObservedRunningTime="2025-10-14 08:06:13.382255799 +0000 UTC m=+906.363495251" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.422203 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.424762 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.425021 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5627s\" (UniqueName: \"kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.425050 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.425111 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.425717 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.426186 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.446643 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.460514 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5627s\" (UniqueName: \"kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s\") pod \"dnsmasq-dns-746b7bb85-2txhh\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.481466 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.486563 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.490002 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.492300 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.565304 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.629085 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.629184 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.629263 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.630028 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.630077 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gsb5\" (UniqueName: \"kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.730172 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1764ec79-d795-4857-a364-1f2326a0291e" path="/var/lib/kubelet/pods/1764ec79-d795-4857-a364-1f2326a0291e/volumes" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.730556 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7fae819-1eef-4b7e-a933-34ba26155759" path="/var/lib/kubelet/pods/d7fae819-1eef-4b7e-a933-34ba26155759/volumes" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.731402 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.731456 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.731489 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.731508 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.731522 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gsb5\" (UniqueName: \"kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.732531 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.732531 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.732600 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.732618 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.748112 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gsb5\" (UniqueName: \"kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5\") pod \"dnsmasq-dns-7764bd4845-bw7l6\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:13 crc kubenswrapper[5002]: I1014 08:06:13.809663 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.331482 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.377253 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2b959df9-bb11-4ee8-8af1-73beac51cfba","Type":"ContainerStarted","Data":"47f49275877844f1976cd6b4e795ab660a532fa78c81e644924f4523e58c8e3f"} Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.379418 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" event={"ID":"3699a03e-0a96-40d8-ac3c-effd67d9729c","Type":"ContainerStarted","Data":"c400863da2804cda8f7e5a91d200ae2c32d335fc9eac8a884bc43fa0d56b5502"} Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.379680 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="dnsmasq-dns" containerID="cri-o://c400863da2804cda8f7e5a91d200ae2c32d335fc9eac8a884bc43fa0d56b5502" gracePeriod=10 Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.398966 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" podStartSLOduration=8.637239084 podStartE2EDuration="17.398950304s" podCreationTimestamp="2025-10-14 08:05:57 +0000 UTC" firstStartedPulling="2025-10-14 08:06:02.369002942 +0000 UTC m=+895.350242404" lastFinishedPulling="2025-10-14 08:06:11.130714172 +0000 UTC m=+904.111953624" observedRunningTime="2025-10-14 08:06:14.396758916 +0000 UTC m=+907.377998388" watchObservedRunningTime="2025-10-14 08:06:14.398950304 +0000 UTC m=+907.380189756" Oct 14 08:06:14 crc kubenswrapper[5002]: W1014 08:06:14.515431 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4fce06a_4448_455d_a569_789da2c47342.slice/crio-12c2764ad79e579d2c68da57d88480af9fc4918842c2e693a1f8ca9641cf552c WatchSource:0}: Error finding container 12c2764ad79e579d2c68da57d88480af9fc4918842c2e693a1f8ca9641cf552c: Status 404 returned error can't find the container with id 12c2764ad79e579d2c68da57d88480af9fc4918842c2e693a1f8ca9641cf552c Oct 14 08:06:14 crc kubenswrapper[5002]: I1014 08:06:14.948102 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-b85t2"] Oct 14 08:06:15 crc kubenswrapper[5002]: I1014 08:06:15.387830 5002 generic.go:334] "Generic (PLEG): container finished" podID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerID="c400863da2804cda8f7e5a91d200ae2c32d335fc9eac8a884bc43fa0d56b5502" exitCode=0 Oct 14 08:06:15 crc kubenswrapper[5002]: I1014 08:06:15.388133 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" event={"ID":"3699a03e-0a96-40d8-ac3c-effd67d9729c","Type":"ContainerDied","Data":"c400863da2804cda8f7e5a91d200ae2c32d335fc9eac8a884bc43fa0d56b5502"} Oct 14 08:06:15 crc kubenswrapper[5002]: I1014 08:06:15.390074 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="dnsmasq-dns" containerID="cri-o://a482acc915c7fdd1cb544a60b8d910573b0396ead0a4ee90e0e85e1de922c709" gracePeriod=10 Oct 14 08:06:15 crc kubenswrapper[5002]: I1014 08:06:15.390303 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" event={"ID":"f4fce06a-4448-455d-a569-789da2c47342","Type":"ContainerStarted","Data":"12c2764ad79e579d2c68da57d88480af9fc4918842c2e693a1f8ca9641cf552c"} Oct 14 08:06:16 crc kubenswrapper[5002]: I1014 08:06:16.403697 5002 generic.go:334] "Generic (PLEG): container finished" podID="190e7b2c-275b-4b93-8485-46659be96268" containerID="a482acc915c7fdd1cb544a60b8d910573b0396ead0a4ee90e0e85e1de922c709" exitCode=0 Oct 14 08:06:16 crc kubenswrapper[5002]: I1014 08:06:16.403747 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" event={"ID":"190e7b2c-275b-4b93-8485-46659be96268","Type":"ContainerDied","Data":"a482acc915c7fdd1cb544a60b8d910573b0396ead0a4ee90e0e85e1de922c709"} Oct 14 08:06:18 crc kubenswrapper[5002]: I1014 08:06:18.532502 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:06:19 crc kubenswrapper[5002]: W1014 08:06:19.727105 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8622c6ea_480a_457b_b42c_2adad7acc425.slice/crio-42d3446c7f1e9993fa686d69be7a601d760dc82b411bed56e24f9fedfbadf55e WatchSource:0}: Error finding container 42d3446c7f1e9993fa686d69be7a601d760dc82b411bed56e24f9fedfbadf55e: Status 404 returned error can't find the container with id 42d3446c7f1e9993fa686d69be7a601d760dc82b411bed56e24f9fedfbadf55e Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.787988 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.944467 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config\") pod \"3699a03e-0a96-40d8-ac3c-effd67d9729c\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.944556 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tww96\" (UniqueName: \"kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96\") pod \"3699a03e-0a96-40d8-ac3c-effd67d9729c\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.944605 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc\") pod \"3699a03e-0a96-40d8-ac3c-effd67d9729c\" (UID: \"3699a03e-0a96-40d8-ac3c-effd67d9729c\") " Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.950976 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96" (OuterVolumeSpecName: "kube-api-access-tww96") pod "3699a03e-0a96-40d8-ac3c-effd67d9729c" (UID: "3699a03e-0a96-40d8-ac3c-effd67d9729c"). InnerVolumeSpecName "kube-api-access-tww96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:19 crc kubenswrapper[5002]: I1014 08:06:19.998194 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config" (OuterVolumeSpecName: "config") pod "3699a03e-0a96-40d8-ac3c-effd67d9729c" (UID: "3699a03e-0a96-40d8-ac3c-effd67d9729c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.012569 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3699a03e-0a96-40d8-ac3c-effd67d9729c" (UID: "3699a03e-0a96-40d8-ac3c-effd67d9729c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.046608 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.046677 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tww96\" (UniqueName: \"kubernetes.io/projected/3699a03e-0a96-40d8-ac3c-effd67d9729c-kube-api-access-tww96\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.046705 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3699a03e-0a96-40d8-ac3c-effd67d9729c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.361513 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.440550 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.441070 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b486d97-dcpvh" event={"ID":"3699a03e-0a96-40d8-ac3c-effd67d9729c","Type":"ContainerDied","Data":"c32ce2244c8ac7f3621401133b86818e0c27f92c6d06b0adfe831f91648195bd"} Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.441137 5002 scope.go:117] "RemoveContainer" containerID="c400863da2804cda8f7e5a91d200ae2c32d335fc9eac8a884bc43fa0d56b5502" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.446133 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" event={"ID":"190e7b2c-275b-4b93-8485-46659be96268","Type":"ContainerDied","Data":"686dd9c40a57d91f290877e05593947acfd293ae831cc59fef7dc07055ee7006"} Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.446254 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.448359 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-b85t2" event={"ID":"8622c6ea-480a-457b-b42c-2adad7acc425","Type":"ContainerStarted","Data":"42d3446c7f1e9993fa686d69be7a601d760dc82b411bed56e24f9fedfbadf55e"} Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.451727 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tm44\" (UniqueName: \"kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44\") pod \"190e7b2c-275b-4b93-8485-46659be96268\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.452373 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config\") pod \"190e7b2c-275b-4b93-8485-46659be96268\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.452456 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc\") pod \"190e7b2c-275b-4b93-8485-46659be96268\" (UID: \"190e7b2c-275b-4b93-8485-46659be96268\") " Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.463731 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44" (OuterVolumeSpecName: "kube-api-access-8tm44") pod "190e7b2c-275b-4b93-8485-46659be96268" (UID: "190e7b2c-275b-4b93-8485-46659be96268"). InnerVolumeSpecName "kube-api-access-8tm44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.491323 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.499759 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586b486d97-dcpvh"] Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.502212 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "190e7b2c-275b-4b93-8485-46659be96268" (UID: "190e7b2c-275b-4b93-8485-46659be96268"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.506561 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config" (OuterVolumeSpecName: "config") pod "190e7b2c-275b-4b93-8485-46659be96268" (UID: "190e7b2c-275b-4b93-8485-46659be96268"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.554934 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tm44\" (UniqueName: \"kubernetes.io/projected/190e7b2c-275b-4b93-8485-46659be96268-kube-api-access-8tm44\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.554973 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.554985 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e7b2c-275b-4b93-8485-46659be96268-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.711158 5002 scope.go:117] "RemoveContainer" containerID="5e9d56ec78f3e463a8b0e9a06bb0372a93e6ddf842e85e4b4067c6ac574fd867" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.829653 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:06:20 crc kubenswrapper[5002]: W1014 08:06:20.851024 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf836312e_c43a_47ed_bfd8_a464948d9687.slice/crio-7095736e637377fbd24a00ba797fcb5e091b3f43082a8fddc267369579980dd2 WatchSource:0}: Error finding container 7095736e637377fbd24a00ba797fcb5e091b3f43082a8fddc267369579980dd2: Status 404 returned error can't find the container with id 7095736e637377fbd24a00ba797fcb5e091b3f43082a8fddc267369579980dd2 Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.863339 5002 scope.go:117] "RemoveContainer" containerID="a482acc915c7fdd1cb544a60b8d910573b0396ead0a4ee90e0e85e1de922c709" Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.929912 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.934038 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-55tk2"] Oct 14 08:06:20 crc kubenswrapper[5002]: I1014 08:06:20.953578 5002 scope.go:117] "RemoveContainer" containerID="3120c99ba207b350798c968c9a2639ebeb0dc1a22a36374d90a8273f9c3e13be" Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.470457 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4fce06a-4448-455d-a569-789da2c47342" containerID="ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac" exitCode=0 Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.470947 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" event={"ID":"f4fce06a-4448-455d-a569-789da2c47342","Type":"ContainerDied","Data":"ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac"} Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.482900 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"355bb6ac-9bd2-411e-81d4-400069911641","Type":"ContainerStarted","Data":"4186fb4872666f0c933430662870afbf2684243f3bce1902460eaafea90ada04"} Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.490576 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" event={"ID":"f836312e-c43a-47ed-bfd8-a464948d9687","Type":"ContainerStarted","Data":"7095736e637377fbd24a00ba797fcb5e091b3f43082a8fddc267369579980dd2"} Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.731289 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="190e7b2c-275b-4b93-8485-46659be96268" path="/var/lib/kubelet/pods/190e7b2c-275b-4b93-8485-46659be96268/volumes" Oct 14 08:06:21 crc kubenswrapper[5002]: I1014 08:06:21.732211 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" path="/var/lib/kubelet/pods/3699a03e-0a96-40d8-ac3c-effd67d9729c/volumes" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.523768 5002 generic.go:334] "Generic (PLEG): container finished" podID="4074edb0-c770-4819-a4bd-581f3e7c6e23" containerID="212d0019f568742015b904ea61485d287994d9c349a0b785be963fdeb3232ea1" exitCode=0 Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.523825 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-n5j7z" event={"ID":"4074edb0-c770-4819-a4bd-581f3e7c6e23","Type":"ContainerDied","Data":"212d0019f568742015b904ea61485d287994d9c349a0b785be963fdeb3232ea1"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.526474 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerStarted","Data":"2336f75047f1dbec2b388320d890aaa908a5b9cf671f7d3dd3b7b2a45675dbaa"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.529777 5002 generic.go:334] "Generic (PLEG): container finished" podID="f836312e-c43a-47ed-bfd8-a464948d9687" containerID="46fe38a4f5bd2fc6926403544eba335c1806269d42a56a01e2fd5c41aab1aa07" exitCode=0 Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.529888 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" event={"ID":"f836312e-c43a-47ed-bfd8-a464948d9687","Type":"ContainerDied","Data":"46fe38a4f5bd2fc6926403544eba335c1806269d42a56a01e2fd5c41aab1aa07"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.534069 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7" event={"ID":"f9439452-e42f-4fa8-ac61-65b5389ff828","Type":"ContainerStarted","Data":"880001bdd8ce50d75b80f6319cef0ed966de30ab1674adad077931f015ea119f"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.534207 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.537997 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d0a98c43-f790-476c-822c-e9b5957bbb78","Type":"ContainerStarted","Data":"a7bf8d09dde8d5ada18f906946b904ea8163ba2badcd902abb4ff35e7c39c905"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.538143 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.540697 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2c1f1d79-cecd-4eda-a876-11dfb83ab76b","Type":"ContainerStarted","Data":"a8f7dc4e446a83667a21d0a38123ee300faa14a1a6f8387928153c352da9b0cf"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.543270 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"705e9cf2-0759-49ed-8407-a73c3aed75cf","Type":"ContainerStarted","Data":"85983ba010f297dda840e3db77042539438e31f585660065321d9334cf249168"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.548384 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2b959df9-bb11-4ee8-8af1-73beac51cfba","Type":"ContainerStarted","Data":"39e2e322035e0aab55953f5bba386c09bd416a3a53ec0c702801e166e99e64a1"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.549870 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fd0a0da-37d6-424c-b8ad-87791d2b155e","Type":"ContainerStarted","Data":"459fafadcc02de6779287720ad9e5ceacb41d3138ae237077a4b199d56698627"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.550079 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.555479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerStarted","Data":"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9"} Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.596051 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=11.642515425 podStartE2EDuration="20.596034489s" podCreationTimestamp="2025-10-14 08:06:02 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.790066489 +0000 UTC m=+904.771305941" lastFinishedPulling="2025-10-14 08:06:20.743585553 +0000 UTC m=+913.724825005" observedRunningTime="2025-10-14 08:06:22.56937143 +0000 UTC m=+915.550610892" watchObservedRunningTime="2025-10-14 08:06:22.596034489 +0000 UTC m=+915.577273941" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.667155 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tw5d7" podStartSLOduration=6.716561085 podStartE2EDuration="15.667134841s" podCreationTimestamp="2025-10-14 08:06:07 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.790000147 +0000 UTC m=+904.771239599" lastFinishedPulling="2025-10-14 08:06:20.740573903 +0000 UTC m=+913.721813355" observedRunningTime="2025-10-14 08:06:22.664425409 +0000 UTC m=+915.645664861" watchObservedRunningTime="2025-10-14 08:06:22.667134841 +0000 UTC m=+915.648374293" Oct 14 08:06:22 crc kubenswrapper[5002]: I1014 08:06:22.705348 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.055187196 podStartE2EDuration="18.705326337s" podCreationTimestamp="2025-10-14 08:06:04 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.785462526 +0000 UTC m=+904.766701978" lastFinishedPulling="2025-10-14 08:06:20.435601667 +0000 UTC m=+913.416841119" observedRunningTime="2025-10-14 08:06:22.700221861 +0000 UTC m=+915.681461303" watchObservedRunningTime="2025-10-14 08:06:22.705326337 +0000 UTC m=+915.686565789" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.491505 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7869c47d6c-55tk2" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.100:5353: i/o timeout" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.563069 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"705e9cf2-0759-49ed-8407-a73c3aed75cf","Type":"ContainerStarted","Data":"01886b56ab6201ec33d5bcbe15540a5011bdf3ae5c21d7668b5f51f050976801"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.566735 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" event={"ID":"f836312e-c43a-47ed-bfd8-a464948d9687","Type":"ContainerStarted","Data":"ab30045065cc39c7fcf0a9122420726490537441cdae94dd465a50fe3e9cd320"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.566787 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.568305 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"2b959df9-bb11-4ee8-8af1-73beac51cfba","Type":"ContainerStarted","Data":"5d5220a1e1897a60d9bea018acdf6ed07eb7f570df61fd4298ff839462e2799e"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.570054 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-b85t2" event={"ID":"8622c6ea-480a-457b-b42c-2adad7acc425","Type":"ContainerStarted","Data":"c81b0c3cf2c83e2e7b07ded3e16d1787576fcae749d3fe5d9fc0da3c908d22c6"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.572027 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" event={"ID":"f4fce06a-4448-455d-a569-789da2c47342","Type":"ContainerStarted","Data":"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.572163 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.573849 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-n5j7z" event={"ID":"4074edb0-c770-4819-a4bd-581f3e7c6e23","Type":"ContainerStarted","Data":"c57696e4183534425b542507ad979827bd1d1ba458babd681788079c0f32a465"} Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.616273 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.5863666309999997 podStartE2EDuration="13.616254099s" podCreationTimestamp="2025-10-14 08:06:10 +0000 UTC" firstStartedPulling="2025-10-14 08:06:13.225461036 +0000 UTC m=+906.206700488" lastFinishedPulling="2025-10-14 08:06:23.255348494 +0000 UTC m=+916.236587956" observedRunningTime="2025-10-14 08:06:23.586803054 +0000 UTC m=+916.568042516" watchObservedRunningTime="2025-10-14 08:06:23.616254099 +0000 UTC m=+916.597493551" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.650778 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" podStartSLOduration=10.650756386 podStartE2EDuration="10.650756386s" podCreationTimestamp="2025-10-14 08:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:06:23.620637345 +0000 UTC m=+916.601876817" watchObservedRunningTime="2025-10-14 08:06:23.650756386 +0000 UTC m=+916.631995848" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.651557 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-b85t2" podStartSLOduration=7.245970211 podStartE2EDuration="10.651551048s" podCreationTimestamp="2025-10-14 08:06:13 +0000 UTC" firstStartedPulling="2025-10-14 08:06:19.836311049 +0000 UTC m=+912.817550501" lastFinishedPulling="2025-10-14 08:06:23.241891886 +0000 UTC m=+916.223131338" observedRunningTime="2025-10-14 08:06:23.640155195 +0000 UTC m=+916.621394657" watchObservedRunningTime="2025-10-14 08:06:23.651551048 +0000 UTC m=+916.632790510" Oct 14 08:06:23 crc kubenswrapper[5002]: I1014 08:06:23.670277 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.76225773 podStartE2EDuration="12.670261885s" podCreationTimestamp="2025-10-14 08:06:11 +0000 UTC" firstStartedPulling="2025-10-14 08:06:13.346070175 +0000 UTC m=+906.327309617" lastFinishedPulling="2025-10-14 08:06:23.25407431 +0000 UTC m=+916.235313772" observedRunningTime="2025-10-14 08:06:23.662811867 +0000 UTC m=+916.644051329" watchObservedRunningTime="2025-10-14 08:06:23.670261885 +0000 UTC m=+916.651501337" Oct 14 08:06:24 crc kubenswrapper[5002]: I1014 08:06:24.245492 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:24 crc kubenswrapper[5002]: I1014 08:06:24.429612 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:24 crc kubenswrapper[5002]: I1014 08:06:24.586276 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-n5j7z" event={"ID":"4074edb0-c770-4819-a4bd-581f3e7c6e23","Type":"ContainerStarted","Data":"9428eb2331cac90c54339f828e7bfac96551271b9c516d93d85e79939a96e04a"} Oct 14 08:06:24 crc kubenswrapper[5002]: I1014 08:06:24.631871 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" podStartSLOduration=11.631822134 podStartE2EDuration="11.631822134s" podCreationTimestamp="2025-10-14 08:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:06:23.692437786 +0000 UTC m=+916.673677238" watchObservedRunningTime="2025-10-14 08:06:24.631822134 +0000 UTC m=+917.613061626" Oct 14 08:06:24 crc kubenswrapper[5002]: I1014 08:06:24.634806 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-n5j7z" podStartSLOduration=8.829956715 podStartE2EDuration="17.634794303s" podCreationTimestamp="2025-10-14 08:06:07 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.90624851 +0000 UTC m=+904.887487962" lastFinishedPulling="2025-10-14 08:06:20.711086098 +0000 UTC m=+913.692325550" observedRunningTime="2025-10-14 08:06:24.621575191 +0000 UTC m=+917.602814683" watchObservedRunningTime="2025-10-14 08:06:24.634794303 +0000 UTC m=+917.616033785" Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.607512 5002 generic.go:334] "Generic (PLEG): container finished" podID="355bb6ac-9bd2-411e-81d4-400069911641" containerID="4186fb4872666f0c933430662870afbf2684243f3bce1902460eaafea90ada04" exitCode=0 Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.607579 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"355bb6ac-9bd2-411e-81d4-400069911641","Type":"ContainerDied","Data":"4186fb4872666f0c933430662870afbf2684243f3bce1902460eaafea90ada04"} Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.611419 5002 generic.go:334] "Generic (PLEG): container finished" podID="2c1f1d79-cecd-4eda-a876-11dfb83ab76b" containerID="a8f7dc4e446a83667a21d0a38123ee300faa14a1a6f8387928153c352da9b0cf" exitCode=0 Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.611538 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2c1f1d79-cecd-4eda-a876-11dfb83ab76b","Type":"ContainerDied","Data":"a8f7dc4e446a83667a21d0a38123ee300faa14a1a6f8387928153c352da9b0cf"} Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.614999 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:25 crc kubenswrapper[5002]: I1014 08:06:25.615602 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:26 crc kubenswrapper[5002]: I1014 08:06:26.626126 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"355bb6ac-9bd2-411e-81d4-400069911641","Type":"ContainerStarted","Data":"e91c0299614cb1dbe4f3dc0a00879fd5a4bf1e9728de22d6de25b4b51b94b424"} Oct 14 08:06:26 crc kubenswrapper[5002]: I1014 08:06:26.630279 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2c1f1d79-cecd-4eda-a876-11dfb83ab76b","Type":"ContainerStarted","Data":"d5dda9a80be91d288dbe4500f6ab45cc09d646e77ade54ee1e19a5ca3c2fe281"} Oct 14 08:06:26 crc kubenswrapper[5002]: I1014 08:06:26.712157 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.515576032 podStartE2EDuration="27.712100893s" podCreationTimestamp="2025-10-14 08:05:59 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.505682981 +0000 UTC m=+904.486922433" lastFinishedPulling="2025-10-14 08:06:20.702207832 +0000 UTC m=+913.683447294" observedRunningTime="2025-10-14 08:06:26.656606736 +0000 UTC m=+919.637846218" watchObservedRunningTime="2025-10-14 08:06:26.712100893 +0000 UTC m=+919.693340385" Oct 14 08:06:26 crc kubenswrapper[5002]: I1014 08:06:26.716236 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.745930842 podStartE2EDuration="25.716223503s" podCreationTimestamp="2025-10-14 08:06:01 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.794177058 +0000 UTC m=+904.775416510" lastFinishedPulling="2025-10-14 08:06:20.764469719 +0000 UTC m=+913.745709171" observedRunningTime="2025-10-14 08:06:26.705072886 +0000 UTC m=+919.686312418" watchObservedRunningTime="2025-10-14 08:06:26.716223503 +0000 UTC m=+919.697462995" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.245457 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.313132 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.430395 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.503664 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.717318 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.717401 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 08:06:27 crc kubenswrapper[5002]: I1014 08:06:27.928057 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.078760 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:06:28 crc kubenswrapper[5002]: E1014 08:06:28.079222 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079243 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: E1014 08:06:28.079255 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="init" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079262 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="init" Oct 14 08:06:28 crc kubenswrapper[5002]: E1014 08:06:28.079292 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="init" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079301 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="init" Oct 14 08:06:28 crc kubenswrapper[5002]: E1014 08:06:28.079317 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079324 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079522 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3699a03e-0a96-40d8-ac3c-effd67d9729c" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.079544 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="190e7b2c-275b-4b93-8485-46659be96268" containerName="dnsmasq-dns" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.080595 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.090540 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.091166 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bf7hk" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.091211 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.091479 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.091520 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.104549 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.104872 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.104960 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.105081 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-config\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.105195 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4sq7\" (UniqueName: \"kubernetes.io/projected/3b8bf477-7229-46cd-aabc-2ae7794a694b-kube-api-access-s4sq7\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.105296 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-scripts\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.105381 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206375 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206736 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206762 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206793 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-config\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206850 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4sq7\" (UniqueName: \"kubernetes.io/projected/3b8bf477-7229-46cd-aabc-2ae7794a694b-kube-api-access-s4sq7\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206895 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-scripts\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.206917 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.207415 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.208163 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-config\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.208226 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8bf477-7229-46cd-aabc-2ae7794a694b-scripts\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.211276 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.212053 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.218484 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8bf477-7229-46cd-aabc-2ae7794a694b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.220900 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4sq7\" (UniqueName: \"kubernetes.io/projected/3b8bf477-7229-46cd-aabc-2ae7794a694b-kube-api-access-s4sq7\") pod \"ovn-northd-0\" (UID: \"3b8bf477-7229-46cd-aabc-2ae7794a694b\") " pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.395705 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.569424 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.844253 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.897103 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.906868 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:28 crc kubenswrapper[5002]: W1014 08:06:28.907106 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8bf477_7229_46cd_aabc_2ae7794a694b.slice/crio-db4d192f5173ee0a789239ed4ad9fdf381755191f496edd6f179cfa64be7b2ad WatchSource:0}: Error finding container db4d192f5173ee0a789239ed4ad9fdf381755191f496edd6f179cfa64be7b2ad: Status 404 returned error can't find the container with id db4d192f5173ee0a789239ed4ad9fdf381755191f496edd6f179cfa64be7b2ad Oct 14 08:06:28 crc kubenswrapper[5002]: I1014 08:06:28.907129 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="dnsmasq-dns" containerID="cri-o://32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87" gracePeriod=10 Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.363593 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.552494 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc\") pod \"f4fce06a-4448-455d-a569-789da2c47342\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.552577 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5627s\" (UniqueName: \"kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s\") pod \"f4fce06a-4448-455d-a569-789da2c47342\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.552646 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb\") pod \"f4fce06a-4448-455d-a569-789da2c47342\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.552723 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config\") pod \"f4fce06a-4448-455d-a569-789da2c47342\" (UID: \"f4fce06a-4448-455d-a569-789da2c47342\") " Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.560710 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s" (OuterVolumeSpecName: "kube-api-access-5627s") pod "f4fce06a-4448-455d-a569-789da2c47342" (UID: "f4fce06a-4448-455d-a569-789da2c47342"). InnerVolumeSpecName "kube-api-access-5627s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.609369 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config" (OuterVolumeSpecName: "config") pod "f4fce06a-4448-455d-a569-789da2c47342" (UID: "f4fce06a-4448-455d-a569-789da2c47342"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.618731 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f4fce06a-4448-455d-a569-789da2c47342" (UID: "f4fce06a-4448-455d-a569-789da2c47342"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.618979 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f4fce06a-4448-455d-a569-789da2c47342" (UID: "f4fce06a-4448-455d-a569-789da2c47342"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.654476 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.654530 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5627s\" (UniqueName: \"kubernetes.io/projected/f4fce06a-4448-455d-a569-789da2c47342-kube-api-access-5627s\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.654628 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.654651 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4fce06a-4448-455d-a569-789da2c47342-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.666047 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4fce06a-4448-455d-a569-789da2c47342" containerID="32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87" exitCode=0 Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.666106 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.666156 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" event={"ID":"f4fce06a-4448-455d-a569-789da2c47342","Type":"ContainerDied","Data":"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87"} Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.666359 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-2txhh" event={"ID":"f4fce06a-4448-455d-a569-789da2c47342","Type":"ContainerDied","Data":"12c2764ad79e579d2c68da57d88480af9fc4918842c2e693a1f8ca9641cf552c"} Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.666439 5002 scope.go:117] "RemoveContainer" containerID="32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.668106 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3b8bf477-7229-46cd-aabc-2ae7794a694b","Type":"ContainerStarted","Data":"db4d192f5173ee0a789239ed4ad9fdf381755191f496edd6f179cfa64be7b2ad"} Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.697711 5002 scope.go:117] "RemoveContainer" containerID="ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.710143 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.715061 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-2txhh"] Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.731016 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4fce06a-4448-455d-a569-789da2c47342" path="/var/lib/kubelet/pods/f4fce06a-4448-455d-a569-789da2c47342/volumes" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.755124 5002 scope.go:117] "RemoveContainer" containerID="32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87" Oct 14 08:06:29 crc kubenswrapper[5002]: E1014 08:06:29.755614 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87\": container with ID starting with 32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87 not found: ID does not exist" containerID="32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.755650 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87"} err="failed to get container status \"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87\": rpc error: code = NotFound desc = could not find container \"32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87\": container with ID starting with 32c3b4e15deab580a9fdb55de9ba9d4452b526e2a5a6f8623e0001fb2874bb87 not found: ID does not exist" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.755677 5002 scope.go:117] "RemoveContainer" containerID="ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac" Oct 14 08:06:29 crc kubenswrapper[5002]: E1014 08:06:29.756281 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac\": container with ID starting with ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac not found: ID does not exist" containerID="ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac" Oct 14 08:06:29 crc kubenswrapper[5002]: I1014 08:06:29.756316 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac"} err="failed to get container status \"ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac\": rpc error: code = NotFound desc = could not find container \"ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac\": container with ID starting with ba2e5d17609f79a256445278a4f3b2a8f086cbff3e888f7cea4ed81b8ea06eac not found: ID does not exist" Oct 14 08:06:31 crc kubenswrapper[5002]: I1014 08:06:31.086861 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 08:06:31 crc kubenswrapper[5002]: I1014 08:06:31.087290 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 08:06:31 crc kubenswrapper[5002]: I1014 08:06:31.692001 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3b8bf477-7229-46cd-aabc-2ae7794a694b","Type":"ContainerStarted","Data":"cf25b48b12fe339006e08341ff3d9a7235884b115def64009f6c1496c08b4c15"} Oct 14 08:06:31 crc kubenswrapper[5002]: I1014 08:06:31.950495 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 08:06:32 crc kubenswrapper[5002]: I1014 08:06:32.007224 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="355bb6ac-9bd2-411e-81d4-400069911641" containerName="galera" probeResult="failure" output=< Oct 14 08:06:32 crc kubenswrapper[5002]: wsrep_local_state_comment (Joined) differs from Synced Oct 14 08:06:32 crc kubenswrapper[5002]: > Oct 14 08:06:32 crc kubenswrapper[5002]: I1014 08:06:32.614080 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:32 crc kubenswrapper[5002]: I1014 08:06:32.614134 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:32 crc kubenswrapper[5002]: I1014 08:06:32.703470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3b8bf477-7229-46cd-aabc-2ae7794a694b","Type":"ContainerStarted","Data":"1368e436d7d7ba13b2824df2317f94ef79d5fc00610c297c1848a27c7e9e7f1d"} Oct 14 08:06:32 crc kubenswrapper[5002]: I1014 08:06:32.730206 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:33 crc kubenswrapper[5002]: I1014 08:06:33.798675 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 08:06:35 crc kubenswrapper[5002]: I1014 08:06:35.024523 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 08:06:35 crc kubenswrapper[5002]: I1014 08:06:35.735182 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 08:06:35 crc kubenswrapper[5002]: I1014 08:06:35.754497 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=6.04753079 podStartE2EDuration="7.754481844s" podCreationTimestamp="2025-10-14 08:06:28 +0000 UTC" firstStartedPulling="2025-10-14 08:06:28.916073023 +0000 UTC m=+921.897312475" lastFinishedPulling="2025-10-14 08:06:30.623024067 +0000 UTC m=+923.604263529" observedRunningTime="2025-10-14 08:06:35.753064046 +0000 UTC m=+928.734303528" watchObservedRunningTime="2025-10-14 08:06:35.754481844 +0000 UTC m=+928.735721306" Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.218035 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.218408 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.218479 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.219496 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.219598 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a" gracePeriod=600 Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.772122 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a" exitCode=0 Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.772193 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a"} Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.772536 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b"} Oct 14 08:06:39 crc kubenswrapper[5002]: I1014 08:06:39.772570 5002 scope.go:117] "RemoveContainer" containerID="d2d32dc4d61563130555e500aa2bc885f7b4a1f42df75a06c1406de4a782182f" Oct 14 08:06:41 crc kubenswrapper[5002]: I1014 08:06:41.169325 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.603160 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-md8d9"] Oct 14 08:06:42 crc kubenswrapper[5002]: E1014 08:06:42.604618 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="dnsmasq-dns" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.604712 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="dnsmasq-dns" Oct 14 08:06:42 crc kubenswrapper[5002]: E1014 08:06:42.604800 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="init" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.604879 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="init" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.605100 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4fce06a-4448-455d-a569-789da2c47342" containerName="dnsmasq-dns" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.605659 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.611038 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-md8d9"] Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.790382 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvwwd\" (UniqueName: \"kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd\") pod \"keystone-db-create-md8d9\" (UID: \"70b9e9fc-6eb1-4948-9f50-c9245db17ea8\") " pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.813988 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vgqc8"] Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.815315 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.823080 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vgqc8"] Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.891562 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvwwd\" (UniqueName: \"kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd\") pod \"keystone-db-create-md8d9\" (UID: \"70b9e9fc-6eb1-4948-9f50-c9245db17ea8\") " pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.931000 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvwwd\" (UniqueName: \"kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd\") pod \"keystone-db-create-md8d9\" (UID: \"70b9e9fc-6eb1-4948-9f50-c9245db17ea8\") " pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:42 crc kubenswrapper[5002]: I1014 08:06:42.993013 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5276\" (UniqueName: \"kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276\") pod \"placement-db-create-vgqc8\" (UID: \"6eb5f6d1-6253-4c86-9953-b83e40c97276\") " pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.094600 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5276\" (UniqueName: \"kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276\") pod \"placement-db-create-vgqc8\" (UID: \"6eb5f6d1-6253-4c86-9953-b83e40c97276\") " pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.125913 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5276\" (UniqueName: \"kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276\") pod \"placement-db-create-vgqc8\" (UID: \"6eb5f6d1-6253-4c86-9953-b83e40c97276\") " pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.141885 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.223636 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.474946 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.532503 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-md8d9"] Oct 14 08:06:43 crc kubenswrapper[5002]: W1014 08:06:43.538389 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70b9e9fc_6eb1_4948_9f50_c9245db17ea8.slice/crio-64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7 WatchSource:0}: Error finding container 64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7: Status 404 returned error can't find the container with id 64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7 Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.632596 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vgqc8"] Oct 14 08:06:43 crc kubenswrapper[5002]: W1014 08:06:43.637616 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb5f6d1_6253_4c86_9953_b83e40c97276.slice/crio-ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf WatchSource:0}: Error finding container ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf: Status 404 returned error can't find the container with id ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.817192 5002 generic.go:334] "Generic (PLEG): container finished" podID="70b9e9fc-6eb1-4948-9f50-c9245db17ea8" containerID="205c0cb1f3dad50fced5285aae956e6710f40ee547061abb544105521a1938d8" exitCode=0 Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.817456 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-md8d9" event={"ID":"70b9e9fc-6eb1-4948-9f50-c9245db17ea8","Type":"ContainerDied","Data":"205c0cb1f3dad50fced5285aae956e6710f40ee547061abb544105521a1938d8"} Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.817512 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-md8d9" event={"ID":"70b9e9fc-6eb1-4948-9f50-c9245db17ea8","Type":"ContainerStarted","Data":"64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7"} Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.820199 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vgqc8" event={"ID":"6eb5f6d1-6253-4c86-9953-b83e40c97276","Type":"ContainerStarted","Data":"b02b0c07dcd5d213197baaac31e1f4092b3042aaa519f3228584e603a71e01c6"} Oct 14 08:06:43 crc kubenswrapper[5002]: I1014 08:06:43.820260 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vgqc8" event={"ID":"6eb5f6d1-6253-4c86-9953-b83e40c97276","Type":"ContainerStarted","Data":"ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf"} Oct 14 08:06:44 crc kubenswrapper[5002]: I1014 08:06:44.835229 5002 generic.go:334] "Generic (PLEG): container finished" podID="6eb5f6d1-6253-4c86-9953-b83e40c97276" containerID="b02b0c07dcd5d213197baaac31e1f4092b3042aaa519f3228584e603a71e01c6" exitCode=0 Oct 14 08:06:44 crc kubenswrapper[5002]: I1014 08:06:44.835299 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vgqc8" event={"ID":"6eb5f6d1-6253-4c86-9953-b83e40c97276","Type":"ContainerDied","Data":"b02b0c07dcd5d213197baaac31e1f4092b3042aaa519f3228584e603a71e01c6"} Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.226173 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.232597 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.332650 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvwwd\" (UniqueName: \"kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd\") pod \"70b9e9fc-6eb1-4948-9f50-c9245db17ea8\" (UID: \"70b9e9fc-6eb1-4948-9f50-c9245db17ea8\") " Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.332782 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5276\" (UniqueName: \"kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276\") pod \"6eb5f6d1-6253-4c86-9953-b83e40c97276\" (UID: \"6eb5f6d1-6253-4c86-9953-b83e40c97276\") " Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.339349 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276" (OuterVolumeSpecName: "kube-api-access-d5276") pod "6eb5f6d1-6253-4c86-9953-b83e40c97276" (UID: "6eb5f6d1-6253-4c86-9953-b83e40c97276"). InnerVolumeSpecName "kube-api-access-d5276". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.339424 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd" (OuterVolumeSpecName: "kube-api-access-nvwwd") pod "70b9e9fc-6eb1-4948-9f50-c9245db17ea8" (UID: "70b9e9fc-6eb1-4948-9f50-c9245db17ea8"). InnerVolumeSpecName "kube-api-access-nvwwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.436863 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5276\" (UniqueName: \"kubernetes.io/projected/6eb5f6d1-6253-4c86-9953-b83e40c97276-kube-api-access-d5276\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.436911 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvwwd\" (UniqueName: \"kubernetes.io/projected/70b9e9fc-6eb1-4948-9f50-c9245db17ea8-kube-api-access-nvwwd\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.852980 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vgqc8" event={"ID":"6eb5f6d1-6253-4c86-9953-b83e40c97276","Type":"ContainerDied","Data":"ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf"} Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.853058 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec7e1ba6e5f183d47ff0a02e4a1698b498f1ce8062aff02b82b40cc9eb2d86cf" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.853132 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vgqc8" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.856559 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-md8d9" event={"ID":"70b9e9fc-6eb1-4948-9f50-c9245db17ea8","Type":"ContainerDied","Data":"64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7"} Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.857221 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64a9ddb87153633e4512fdcdef456396dab5d0f3e3fcaf9e09824ba8018f89b7" Oct 14 08:06:45 crc kubenswrapper[5002]: I1014 08:06:45.856718 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-md8d9" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.106101 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rhqxc"] Oct 14 08:06:48 crc kubenswrapper[5002]: E1014 08:06:48.106611 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70b9e9fc-6eb1-4948-9f50-c9245db17ea8" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.106631 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="70b9e9fc-6eb1-4948-9f50-c9245db17ea8" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: E1014 08:06:48.106672 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb5f6d1-6253-4c86-9953-b83e40c97276" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.106683 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb5f6d1-6253-4c86-9953-b83e40c97276" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.106996 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb5f6d1-6253-4c86-9953-b83e40c97276" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.107023 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="70b9e9fc-6eb1-4948-9f50-c9245db17ea8" containerName="mariadb-database-create" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.107763 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.130242 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rhqxc"] Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.297631 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd9ss\" (UniqueName: \"kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss\") pod \"glance-db-create-rhqxc\" (UID: \"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6\") " pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.399509 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd9ss\" (UniqueName: \"kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss\") pod \"glance-db-create-rhqxc\" (UID: \"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6\") " pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.426688 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd9ss\" (UniqueName: \"kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss\") pod \"glance-db-create-rhqxc\" (UID: \"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6\") " pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.436339 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:48 crc kubenswrapper[5002]: I1014 08:06:48.884830 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rhqxc"] Oct 14 08:06:48 crc kubenswrapper[5002]: W1014 08:06:48.891234 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfb1ba9e_3684_4c6a_a33a_82e316a44cd6.slice/crio-5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e WatchSource:0}: Error finding container 5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e: Status 404 returned error can't find the container with id 5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e Oct 14 08:06:49 crc kubenswrapper[5002]: I1014 08:06:49.901935 5002 generic.go:334] "Generic (PLEG): container finished" podID="cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" containerID="9ae5d2a808b5331b090ba5eefd901689444a0e09a5779f1e1a9ff29e4c63e2f1" exitCode=0 Oct 14 08:06:49 crc kubenswrapper[5002]: I1014 08:06:49.902045 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rhqxc" event={"ID":"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6","Type":"ContainerDied","Data":"9ae5d2a808b5331b090ba5eefd901689444a0e09a5779f1e1a9ff29e4c63e2f1"} Oct 14 08:06:49 crc kubenswrapper[5002]: I1014 08:06:49.902353 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rhqxc" event={"ID":"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6","Type":"ContainerStarted","Data":"5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e"} Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.271640 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.473048 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd9ss\" (UniqueName: \"kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss\") pod \"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6\" (UID: \"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6\") " Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.482210 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss" (OuterVolumeSpecName: "kube-api-access-pd9ss") pod "cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" (UID: "cfb1ba9e-3684-4c6a-a33a-82e316a44cd6"). InnerVolumeSpecName "kube-api-access-pd9ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.575363 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd9ss\" (UniqueName: \"kubernetes.io/projected/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6-kube-api-access-pd9ss\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.927581 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rhqxc" event={"ID":"cfb1ba9e-3684-4c6a-a33a-82e316a44cd6","Type":"ContainerDied","Data":"5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e"} Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.927697 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b361b573cbd34fb1e43178da06deac1d8ccb7af0e552a8068a23aa168b9857e" Oct 14 08:06:51 crc kubenswrapper[5002]: I1014 08:06:51.927747 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rhqxc" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.672166 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-a057-account-create-rgqxv"] Oct 14 08:06:52 crc kubenswrapper[5002]: E1014 08:06:52.672722 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" containerName="mariadb-database-create" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.672748 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" containerName="mariadb-database-create" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.673137 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" containerName="mariadb-database-create" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.674116 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.677143 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.680407 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-a057-account-create-rgqxv"] Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.699045 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp64x\" (UniqueName: \"kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x\") pod \"keystone-a057-account-create-rgqxv\" (UID: \"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9\") " pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.802431 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp64x\" (UniqueName: \"kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x\") pod \"keystone-a057-account-create-rgqxv\" (UID: \"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9\") " pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.832885 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp64x\" (UniqueName: \"kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x\") pod \"keystone-a057-account-create-rgqxv\" (UID: \"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9\") " pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.848266 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-44d2-account-create-2k92k"] Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.849347 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.852382 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.863793 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-44d2-account-create-2k92k"] Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.904812 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svj2j\" (UniqueName: \"kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j\") pod \"placement-44d2-account-create-2k92k\" (UID: \"cc62cd95-ad2f-4f14-820f-8545fb48080b\") " pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:52 crc kubenswrapper[5002]: I1014 08:06:52.998748 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.006608 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svj2j\" (UniqueName: \"kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j\") pod \"placement-44d2-account-create-2k92k\" (UID: \"cc62cd95-ad2f-4f14-820f-8545fb48080b\") " pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.028672 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svj2j\" (UniqueName: \"kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j\") pod \"placement-44d2-account-create-2k92k\" (UID: \"cc62cd95-ad2f-4f14-820f-8545fb48080b\") " pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.213650 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.221963 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-a057-account-create-rgqxv"] Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.324414 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.328258 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tw5d7" podUID="f9439452-e42f-4fa8-ac61-65b5389ff828" containerName="ovn-controller" probeResult="failure" output=< Oct 14 08:06:53 crc kubenswrapper[5002]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 08:06:53 crc kubenswrapper[5002]: > Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.328498 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-n5j7z" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.558851 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tw5d7-config-2mx2t"] Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.560496 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.564260 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.577604 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7-config-2mx2t"] Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617019 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617418 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617456 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617516 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzzr9\" (UniqueName: \"kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617740 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.617814 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.656497 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-44d2-account-create-2k92k"] Oct 14 08:06:53 crc kubenswrapper[5002]: W1014 08:06:53.662507 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc62cd95_ad2f_4f14_820f_8545fb48080b.slice/crio-a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d WatchSource:0}: Error finding container a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d: Status 404 returned error can't find the container with id a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718455 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718522 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718581 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718603 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718644 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718695 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzzr9\" (UniqueName: \"kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718961 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.718990 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.719095 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.719611 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.720582 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.738870 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzzr9\" (UniqueName: \"kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9\") pod \"ovn-controller-tw5d7-config-2mx2t\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.881485 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.940824 5002 generic.go:334] "Generic (PLEG): container finished" podID="a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" containerID="e253ff5b25f1f40c213293e16d64f22a7c4b29da53dd2f51fcb96a91b1ea18f7" exitCode=0 Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.940891 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a057-account-create-rgqxv" event={"ID":"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9","Type":"ContainerDied","Data":"e253ff5b25f1f40c213293e16d64f22a7c4b29da53dd2f51fcb96a91b1ea18f7"} Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.940913 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a057-account-create-rgqxv" event={"ID":"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9","Type":"ContainerStarted","Data":"0a6f1e916fba07e7763984c365eeb0fdfb5347a3de96e8b00d4056e76fa2ebb8"} Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.942060 5002 generic.go:334] "Generic (PLEG): container finished" podID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerID="1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9" exitCode=0 Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.942095 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerDied","Data":"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9"} Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.950993 5002 generic.go:334] "Generic (PLEG): container finished" podID="cc62cd95-ad2f-4f14-820f-8545fb48080b" containerID="cae03b9ef2c61b476773b1d165be3ff50a774aa17a2b9d3c4f92dae7e0f7002a" exitCode=0 Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.951057 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-44d2-account-create-2k92k" event={"ID":"cc62cd95-ad2f-4f14-820f-8545fb48080b","Type":"ContainerDied","Data":"cae03b9ef2c61b476773b1d165be3ff50a774aa17a2b9d3c4f92dae7e0f7002a"} Oct 14 08:06:53 crc kubenswrapper[5002]: I1014 08:06:53.951080 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-44d2-account-create-2k92k" event={"ID":"cc62cd95-ad2f-4f14-820f-8545fb48080b","Type":"ContainerStarted","Data":"a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d"} Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.338422 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7-config-2mx2t"] Oct 14 08:06:54 crc kubenswrapper[5002]: W1014 08:06:54.353153 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod583adea6_73c3_4907_a55b_b5b1556e76e0.slice/crio-1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80 WatchSource:0}: Error finding container 1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80: Status 404 returned error can't find the container with id 1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80 Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.959284 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerStarted","Data":"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be"} Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.960024 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.964979 5002 generic.go:334] "Generic (PLEG): container finished" podID="583adea6-73c3-4907-a55b-b5b1556e76e0" containerID="6288dc12ae757ad9e3b98e3c0c7ec5840048d2e57bcc9f727475073969ded676" exitCode=0 Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.965102 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-2mx2t" event={"ID":"583adea6-73c3-4907-a55b-b5b1556e76e0","Type":"ContainerDied","Data":"6288dc12ae757ad9e3b98e3c0c7ec5840048d2e57bcc9f727475073969ded676"} Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.965124 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-2mx2t" event={"ID":"583adea6-73c3-4907-a55b-b5b1556e76e0","Type":"ContainerStarted","Data":"1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80"} Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.966719 5002 generic.go:334] "Generic (PLEG): container finished" podID="f7fe047a-78a9-4abf-a948-167ba509729c" containerID="2336f75047f1dbec2b388320d890aaa908a5b9cf671f7d3dd3b7b2a45675dbaa" exitCode=0 Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.966770 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerDied","Data":"2336f75047f1dbec2b388320d890aaa908a5b9cf671f7d3dd3b7b2a45675dbaa"} Oct 14 08:06:54 crc kubenswrapper[5002]: I1014 08:06:54.986147 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=47.736273142 podStartE2EDuration="56.986131153s" podCreationTimestamp="2025-10-14 08:05:58 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.493300551 +0000 UTC m=+904.474540003" lastFinishedPulling="2025-10-14 08:06:20.743158572 +0000 UTC m=+913.724398014" observedRunningTime="2025-10-14 08:06:54.985346212 +0000 UTC m=+947.966585674" watchObservedRunningTime="2025-10-14 08:06:54.986131153 +0000 UTC m=+947.967370615" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.240741 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.296800 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.340979 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp64x\" (UniqueName: \"kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x\") pod \"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9\" (UID: \"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9\") " Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.344247 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x" (OuterVolumeSpecName: "kube-api-access-wp64x") pod "a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" (UID: "a4e8168d-4cbe-4032-9ab8-02a8c072a5b9"). InnerVolumeSpecName "kube-api-access-wp64x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.442109 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svj2j\" (UniqueName: \"kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j\") pod \"cc62cd95-ad2f-4f14-820f-8545fb48080b\" (UID: \"cc62cd95-ad2f-4f14-820f-8545fb48080b\") " Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.442708 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp64x\" (UniqueName: \"kubernetes.io/projected/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9-kube-api-access-wp64x\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.447133 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j" (OuterVolumeSpecName: "kube-api-access-svj2j") pod "cc62cd95-ad2f-4f14-820f-8545fb48080b" (UID: "cc62cd95-ad2f-4f14-820f-8545fb48080b"). InnerVolumeSpecName "kube-api-access-svj2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.544454 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svj2j\" (UniqueName: \"kubernetes.io/projected/cc62cd95-ad2f-4f14-820f-8545fb48080b-kube-api-access-svj2j\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.977571 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-a057-account-create-rgqxv" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.977585 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-a057-account-create-rgqxv" event={"ID":"a4e8168d-4cbe-4032-9ab8-02a8c072a5b9","Type":"ContainerDied","Data":"0a6f1e916fba07e7763984c365eeb0fdfb5347a3de96e8b00d4056e76fa2ebb8"} Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.977634 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a6f1e916fba07e7763984c365eeb0fdfb5347a3de96e8b00d4056e76fa2ebb8" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.979734 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerStarted","Data":"9efc43d1e74ae820fb9f25e41c46693c27be9380501b3fde947346eda7a623f4"} Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.981210 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.983890 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-44d2-account-create-2k92k" event={"ID":"cc62cd95-ad2f-4f14-820f-8545fb48080b","Type":"ContainerDied","Data":"a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d"} Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.983940 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0fbde3eebbd6842745755afa9688b41528bdaa0ea29776c19746b24831de64d" Oct 14 08:06:55 crc kubenswrapper[5002]: I1014 08:06:55.984020 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-44d2-account-create-2k92k" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.021057 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.952181908 podStartE2EDuration="58.021031102s" podCreationTimestamp="2025-10-14 08:05:58 +0000 UTC" firstStartedPulling="2025-10-14 08:06:11.633240985 +0000 UTC m=+904.614480437" lastFinishedPulling="2025-10-14 08:06:20.702090179 +0000 UTC m=+913.683329631" observedRunningTime="2025-10-14 08:06:56.010738979 +0000 UTC m=+948.991978481" watchObservedRunningTime="2025-10-14 08:06:56.021031102 +0000 UTC m=+949.002270594" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.375084 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459057 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459100 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459128 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459216 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzzr9\" (UniqueName: \"kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459263 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.459300 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts\") pod \"583adea6-73c3-4907-a55b-b5b1556e76e0\" (UID: \"583adea6-73c3-4907-a55b-b5b1556e76e0\") " Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.460185 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.460225 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run" (OuterVolumeSpecName: "var-run") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.460240 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.460782 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts" (OuterVolumeSpecName: "scripts") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.461221 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.475955 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9" (OuterVolumeSpecName: "kube-api-access-fzzr9") pod "583adea6-73c3-4907-a55b-b5b1556e76e0" (UID: "583adea6-73c3-4907-a55b-b5b1556e76e0"). InnerVolumeSpecName "kube-api-access-fzzr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561339 5002 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561373 5002 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561382 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561392 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzzr9\" (UniqueName: \"kubernetes.io/projected/583adea6-73c3-4907-a55b-b5b1556e76e0-kube-api-access-fzzr9\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561401 5002 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/583adea6-73c3-4907-a55b-b5b1556e76e0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.561413 5002 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/583adea6-73c3-4907-a55b-b5b1556e76e0-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.997691 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-2mx2t" Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.997687 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-2mx2t" event={"ID":"583adea6-73c3-4907-a55b-b5b1556e76e0","Type":"ContainerDied","Data":"1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80"} Oct 14 08:06:56 crc kubenswrapper[5002]: I1014 08:06:56.997899 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1826a2a056194ffc4e2a66ad2e70a6252f4f8f08e779ac4e5b0cd58087ff1d80" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.508386 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tw5d7-config-2mx2t"] Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.515242 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tw5d7-config-2mx2t"] Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.628532 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tw5d7-config-9sdlc"] Oct 14 08:06:57 crc kubenswrapper[5002]: E1014 08:06:57.629340 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629390 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: E1014 08:06:57.629428 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="583adea6-73c3-4907-a55b-b5b1556e76e0" containerName="ovn-config" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629450 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="583adea6-73c3-4907-a55b-b5b1556e76e0" containerName="ovn-config" Oct 14 08:06:57 crc kubenswrapper[5002]: E1014 08:06:57.629497 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc62cd95-ad2f-4f14-820f-8545fb48080b" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629517 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc62cd95-ad2f-4f14-820f-8545fb48080b" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629876 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="583adea6-73c3-4907-a55b-b5b1556e76e0" containerName="ovn-config" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629905 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc62cd95-ad2f-4f14-820f-8545fb48080b" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.629926 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" containerName="mariadb-account-create" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.630941 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.634890 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.635239 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7-config-9sdlc"] Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.741629 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="583adea6-73c3-4907-a55b-b5b1556e76e0" path="/var/lib/kubelet/pods/583adea6-73c3-4907-a55b-b5b1556e76e0/volumes" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782378 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782423 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j2gf\" (UniqueName: \"kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782481 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782497 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782698 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.782891 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884611 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884669 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j2gf\" (UniqueName: \"kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884717 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884740 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884797 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.884887 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.885002 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.885012 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.885089 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.885628 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.886591 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.914859 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j2gf\" (UniqueName: \"kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf\") pod \"ovn-controller-tw5d7-config-9sdlc\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:57 crc kubenswrapper[5002]: I1014 08:06:57.953558 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.214322 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tw5d7-config-9sdlc"] Oct 14 08:06:58 crc kubenswrapper[5002]: W1014 08:06:58.220547 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61ba5c41_5492_4003_9778_6271b2d0364f.slice/crio-ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24 WatchSource:0}: Error finding container ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24: Status 404 returned error can't find the container with id ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24 Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.245936 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f8d8-account-create-f78dk"] Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.247330 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.256291 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.258978 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f8d8-account-create-f78dk"] Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.297799 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tw5d7" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.397824 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnwdj\" (UniqueName: \"kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj\") pod \"glance-f8d8-account-create-f78dk\" (UID: \"7ed027b3-8a30-4fee-84c0-70e03887479a\") " pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.499224 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnwdj\" (UniqueName: \"kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj\") pod \"glance-f8d8-account-create-f78dk\" (UID: \"7ed027b3-8a30-4fee-84c0-70e03887479a\") " pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.520637 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnwdj\" (UniqueName: \"kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj\") pod \"glance-f8d8-account-create-f78dk\" (UID: \"7ed027b3-8a30-4fee-84c0-70e03887479a\") " pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.580130 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:06:58 crc kubenswrapper[5002]: I1014 08:06:58.866588 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f8d8-account-create-f78dk"] Oct 14 08:06:59 crc kubenswrapper[5002]: I1014 08:06:59.018697 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f8d8-account-create-f78dk" event={"ID":"7ed027b3-8a30-4fee-84c0-70e03887479a","Type":"ContainerStarted","Data":"2dbb4fe27ddab8efffcce01bd57fd60147a06317ffaee916d742177395b1f8ea"} Oct 14 08:06:59 crc kubenswrapper[5002]: I1014 08:06:59.020439 5002 generic.go:334] "Generic (PLEG): container finished" podID="61ba5c41-5492-4003-9778-6271b2d0364f" containerID="5c6cde16a988f75021402f33ef371619f677ef4c8da9bcb8df53c1c0b1bf63c2" exitCode=0 Oct 14 08:06:59 crc kubenswrapper[5002]: I1014 08:06:59.020471 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-9sdlc" event={"ID":"61ba5c41-5492-4003-9778-6271b2d0364f","Type":"ContainerDied","Data":"5c6cde16a988f75021402f33ef371619f677ef4c8da9bcb8df53c1c0b1bf63c2"} Oct 14 08:06:59 crc kubenswrapper[5002]: I1014 08:06:59.020489 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-9sdlc" event={"ID":"61ba5c41-5492-4003-9778-6271b2d0364f","Type":"ContainerStarted","Data":"ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24"} Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.031373 5002 generic.go:334] "Generic (PLEG): container finished" podID="7ed027b3-8a30-4fee-84c0-70e03887479a" containerID="9eb968117a909bce92185225c7db43136e1ea202e8aa533835cde343baf64a66" exitCode=0 Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.031437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f8d8-account-create-f78dk" event={"ID":"7ed027b3-8a30-4fee-84c0-70e03887479a","Type":"ContainerDied","Data":"9eb968117a909bce92185225c7db43136e1ea202e8aa533835cde343baf64a66"} Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.331016 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458433 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458527 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458592 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j2gf\" (UniqueName: \"kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458621 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458642 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458687 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458719 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn\") pod \"61ba5c41-5492-4003-9778-6271b2d0364f\" (UID: \"61ba5c41-5492-4003-9778-6271b2d0364f\") " Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458738 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run" (OuterVolumeSpecName: "var-run") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.458831 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.459052 5002 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.459068 5002 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.459077 5002 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/61ba5c41-5492-4003-9778-6271b2d0364f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.460119 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts" (OuterVolumeSpecName: "scripts") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.460983 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.478131 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf" (OuterVolumeSpecName: "kube-api-access-8j2gf") pod "61ba5c41-5492-4003-9778-6271b2d0364f" (UID: "61ba5c41-5492-4003-9778-6271b2d0364f"). InnerVolumeSpecName "kube-api-access-8j2gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.560609 5002 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.560663 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j2gf\" (UniqueName: \"kubernetes.io/projected/61ba5c41-5492-4003-9778-6271b2d0364f-kube-api-access-8j2gf\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:00 crc kubenswrapper[5002]: I1014 08:07:00.560680 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ba5c41-5492-4003-9778-6271b2d0364f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.045710 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tw5d7-config-9sdlc" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.048992 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tw5d7-config-9sdlc" event={"ID":"61ba5c41-5492-4003-9778-6271b2d0364f","Type":"ContainerDied","Data":"ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24"} Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.049202 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca8ef9ac1f37a3005d00cad868a4d58038dffe0c1361cfa8451a23c515327b24" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.409004 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tw5d7-config-9sdlc"] Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.420805 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tw5d7-config-9sdlc"] Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.438573 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.579664 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnwdj\" (UniqueName: \"kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj\") pod \"7ed027b3-8a30-4fee-84c0-70e03887479a\" (UID: \"7ed027b3-8a30-4fee-84c0-70e03887479a\") " Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.585500 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj" (OuterVolumeSpecName: "kube-api-access-gnwdj") pod "7ed027b3-8a30-4fee-84c0-70e03887479a" (UID: "7ed027b3-8a30-4fee-84c0-70e03887479a"). InnerVolumeSpecName "kube-api-access-gnwdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.681765 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnwdj\" (UniqueName: \"kubernetes.io/projected/7ed027b3-8a30-4fee-84c0-70e03887479a-kube-api-access-gnwdj\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:01 crc kubenswrapper[5002]: I1014 08:07:01.734251 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61ba5c41-5492-4003-9778-6271b2d0364f" path="/var/lib/kubelet/pods/61ba5c41-5492-4003-9778-6271b2d0364f/volumes" Oct 14 08:07:02 crc kubenswrapper[5002]: I1014 08:07:02.058996 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f8d8-account-create-f78dk" event={"ID":"7ed027b3-8a30-4fee-84c0-70e03887479a","Type":"ContainerDied","Data":"2dbb4fe27ddab8efffcce01bd57fd60147a06317ffaee916d742177395b1f8ea"} Oct 14 08:07:02 crc kubenswrapper[5002]: I1014 08:07:02.059054 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dbb4fe27ddab8efffcce01bd57fd60147a06317ffaee916d742177395b1f8ea" Oct 14 08:07:02 crc kubenswrapper[5002]: I1014 08:07:02.060135 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f8d8-account-create-f78dk" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.465601 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-z4thp"] Oct 14 08:07:03 crc kubenswrapper[5002]: E1014 08:07:03.466860 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed027b3-8a30-4fee-84c0-70e03887479a" containerName="mariadb-account-create" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.466877 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed027b3-8a30-4fee-84c0-70e03887479a" containerName="mariadb-account-create" Oct 14 08:07:03 crc kubenswrapper[5002]: E1014 08:07:03.466913 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61ba5c41-5492-4003-9778-6271b2d0364f" containerName="ovn-config" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.466921 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="61ba5c41-5492-4003-9778-6271b2d0364f" containerName="ovn-config" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.467138 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed027b3-8a30-4fee-84c0-70e03887479a" containerName="mariadb-account-create" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.467189 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="61ba5c41-5492-4003-9778-6271b2d0364f" containerName="ovn-config" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.467799 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.470619 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xvzng" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.471559 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.477070 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-z4thp"] Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.543317 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.543386 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8qqv\" (UniqueName: \"kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.543455 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.543492 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.644772 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.644873 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.644950 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.645017 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8qqv\" (UniqueName: \"kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.651667 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.653528 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.655474 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.660269 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8qqv\" (UniqueName: \"kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv\") pod \"glance-db-sync-z4thp\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:03 crc kubenswrapper[5002]: I1014 08:07:03.795375 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:04 crc kubenswrapper[5002]: I1014 08:07:04.204568 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-z4thp"] Oct 14 08:07:05 crc kubenswrapper[5002]: I1014 08:07:05.097931 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z4thp" event={"ID":"55b80ca7-5604-4c84-96f5-1279ecf2abe7","Type":"ContainerStarted","Data":"d96c9f2281e39628e5d1e17ecfacaefb7c8934a4bbf55b39ed5fe62f497df524"} Oct 14 08:07:09 crc kubenswrapper[5002]: I1014 08:07:09.368064 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:07:09 crc kubenswrapper[5002]: I1014 08:07:09.626202 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.068446 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-snvl8"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.069589 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.081330 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-snvl8"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.167612 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-b87gc"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.173378 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.199165 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b87gc"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.217856 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv98v\" (UniqueName: \"kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v\") pod \"cinder-db-create-snvl8\" (UID: \"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b\") " pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.319151 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46pjs\" (UniqueName: \"kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs\") pod \"barbican-db-create-b87gc\" (UID: \"e29da0d6-d597-4318-ba7f-da9c4d32e518\") " pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.319223 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv98v\" (UniqueName: \"kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v\") pod \"cinder-db-create-snvl8\" (UID: \"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b\") " pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.369462 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv98v\" (UniqueName: \"kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v\") pod \"cinder-db-create-snvl8\" (UID: \"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b\") " pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.384517 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-l8s85"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.385450 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.409354 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l8s85"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.420849 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46pjs\" (UniqueName: \"kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs\") pod \"barbican-db-create-b87gc\" (UID: \"e29da0d6-d597-4318-ba7f-da9c4d32e518\") " pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.438418 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46pjs\" (UniqueName: \"kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs\") pod \"barbican-db-create-b87gc\" (UID: \"e29da0d6-d597-4318-ba7f-da9c4d32e518\") " pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.439220 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.461182 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4tkcz"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.462631 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.466173 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jps45" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.466328 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.466464 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.466648 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.476434 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4tkcz"] Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.492016 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.522709 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.522775 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcdcw\" (UniqueName: \"kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.522800 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvxwj\" (UniqueName: \"kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj\") pod \"neutron-db-create-l8s85\" (UID: \"a3cb8568-04da-484f-a4f8-f1b2e8ab755c\") " pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.522823 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.625110 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.625250 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcdcw\" (UniqueName: \"kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.625277 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvxwj\" (UniqueName: \"kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj\") pod \"neutron-db-create-l8s85\" (UID: \"a3cb8568-04da-484f-a4f8-f1b2e8ab755c\") " pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.625303 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.629058 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.665412 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcdcw\" (UniqueName: \"kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.665419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvxwj\" (UniqueName: \"kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj\") pod \"neutron-db-create-l8s85\" (UID: \"a3cb8568-04da-484f-a4f8-f1b2e8ab755c\") " pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.665615 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle\") pod \"keystone-db-sync-4tkcz\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.722353 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:11 crc kubenswrapper[5002]: I1014 08:07:11.789954 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:15 crc kubenswrapper[5002]: W1014 08:07:15.801491 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa02e7c2_896a_4750_968d_9a4543651928.slice/crio-b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f WatchSource:0}: Error finding container b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f: Status 404 returned error can't find the container with id b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f Oct 14 08:07:15 crc kubenswrapper[5002]: I1014 08:07:15.801813 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4tkcz"] Oct 14 08:07:15 crc kubenswrapper[5002]: I1014 08:07:15.807787 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-snvl8"] Oct 14 08:07:15 crc kubenswrapper[5002]: W1014 08:07:15.809812 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f298eb8_cf82_4d09_93a2_f70fd6e50b9b.slice/crio-4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847 WatchSource:0}: Error finding container 4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847: Status 404 returned error can't find the container with id 4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847 Oct 14 08:07:15 crc kubenswrapper[5002]: I1014 08:07:15.865066 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l8s85"] Oct 14 08:07:15 crc kubenswrapper[5002]: W1014 08:07:15.874831 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3cb8568_04da_484f_a4f8_f1b2e8ab755c.slice/crio-612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9 WatchSource:0}: Error finding container 612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9: Status 404 returned error can't find the container with id 612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9 Oct 14 08:07:15 crc kubenswrapper[5002]: I1014 08:07:15.918808 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b87gc"] Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.193794 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4tkcz" event={"ID":"aa02e7c2-896a-4750-968d-9a4543651928","Type":"ContainerStarted","Data":"b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.196175 5002 generic.go:334] "Generic (PLEG): container finished" podID="9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" containerID="a8368296357610b5e1d057d3b522f52da842ecb976c1b288052d100593fd4fa3" exitCode=0 Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.196240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-snvl8" event={"ID":"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b","Type":"ContainerDied","Data":"a8368296357610b5e1d057d3b522f52da842ecb976c1b288052d100593fd4fa3"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.196265 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-snvl8" event={"ID":"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b","Type":"ContainerStarted","Data":"4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.198112 5002 generic.go:334] "Generic (PLEG): container finished" podID="a3cb8568-04da-484f-a4f8-f1b2e8ab755c" containerID="8fffffe2126815d7948a951b571e0c52ce80cc61ddcbbb78171ac5fc8cbd8be2" exitCode=0 Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.198207 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l8s85" event={"ID":"a3cb8568-04da-484f-a4f8-f1b2e8ab755c","Type":"ContainerDied","Data":"8fffffe2126815d7948a951b571e0c52ce80cc61ddcbbb78171ac5fc8cbd8be2"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.198225 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l8s85" event={"ID":"a3cb8568-04da-484f-a4f8-f1b2e8ab755c","Type":"ContainerStarted","Data":"612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.200532 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b87gc" event={"ID":"e29da0d6-d597-4318-ba7f-da9c4d32e518","Type":"ContainerStarted","Data":"af0ab14336c65a3f322e138fdb595b51a6a72834c4fc62fadcc3868484f6952a"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.202680 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z4thp" event={"ID":"55b80ca7-5604-4c84-96f5-1279ecf2abe7","Type":"ContainerStarted","Data":"8f78daae4f3b6f9db13c31b5682eee751191639b981f868ce72155de15da63ac"} Oct 14 08:07:16 crc kubenswrapper[5002]: I1014 08:07:16.235479 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-z4thp" podStartSLOduration=2.036850055 podStartE2EDuration="13.235459794s" podCreationTimestamp="2025-10-14 08:07:03 +0000 UTC" firstStartedPulling="2025-10-14 08:07:04.21559169 +0000 UTC m=+957.196831142" lastFinishedPulling="2025-10-14 08:07:15.414201419 +0000 UTC m=+968.395440881" observedRunningTime="2025-10-14 08:07:16.232413943 +0000 UTC m=+969.213653405" watchObservedRunningTime="2025-10-14 08:07:16.235459794 +0000 UTC m=+969.216699256" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.216127 5002 generic.go:334] "Generic (PLEG): container finished" podID="e29da0d6-d597-4318-ba7f-da9c4d32e518" containerID="a74e90ead29362576d8f1a8c92b4414432d8a5911887362030cc101440dd71a6" exitCode=0 Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.217069 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b87gc" event={"ID":"e29da0d6-d597-4318-ba7f-da9c4d32e518","Type":"ContainerDied","Data":"a74e90ead29362576d8f1a8c92b4414432d8a5911887362030cc101440dd71a6"} Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.612426 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.626154 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.735106 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv98v\" (UniqueName: \"kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v\") pod \"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b\" (UID: \"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b\") " Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.735170 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvxwj\" (UniqueName: \"kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj\") pod \"a3cb8568-04da-484f-a4f8-f1b2e8ab755c\" (UID: \"a3cb8568-04da-484f-a4f8-f1b2e8ab755c\") " Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.741413 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj" (OuterVolumeSpecName: "kube-api-access-xvxwj") pod "a3cb8568-04da-484f-a4f8-f1b2e8ab755c" (UID: "a3cb8568-04da-484f-a4f8-f1b2e8ab755c"). InnerVolumeSpecName "kube-api-access-xvxwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.743984 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v" (OuterVolumeSpecName: "kube-api-access-bv98v") pod "9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" (UID: "9f298eb8-cf82-4d09-93a2-f70fd6e50b9b"). InnerVolumeSpecName "kube-api-access-bv98v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.836797 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv98v\" (UniqueName: \"kubernetes.io/projected/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b-kube-api-access-bv98v\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:17 crc kubenswrapper[5002]: I1014 08:07:17.836832 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvxwj\" (UniqueName: \"kubernetes.io/projected/a3cb8568-04da-484f-a4f8-f1b2e8ab755c-kube-api-access-xvxwj\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.226983 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-snvl8" event={"ID":"9f298eb8-cf82-4d09-93a2-f70fd6e50b9b","Type":"ContainerDied","Data":"4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847"} Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.227032 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb7c2f9efa5003d9ddc7cf4b3e2042521c7cdb634897f1dd36a49e0538b6847" Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.227100 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-snvl8" Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.233477 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l8s85" Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.234129 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l8s85" event={"ID":"a3cb8568-04da-484f-a4f8-f1b2e8ab755c","Type":"ContainerDied","Data":"612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9"} Oct 14 08:07:18 crc kubenswrapper[5002]: I1014 08:07:18.234312 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="612735fe30b750dc79c1d9cc02b6bc04d5d3b208f6c6ad4c9b565271739aedc9" Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.250667 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b87gc" event={"ID":"e29da0d6-d597-4318-ba7f-da9c4d32e518","Type":"ContainerDied","Data":"af0ab14336c65a3f322e138fdb595b51a6a72834c4fc62fadcc3868484f6952a"} Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.251068 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af0ab14336c65a3f322e138fdb595b51a6a72834c4fc62fadcc3868484f6952a" Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.423723 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.482197 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46pjs\" (UniqueName: \"kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs\") pod \"e29da0d6-d597-4318-ba7f-da9c4d32e518\" (UID: \"e29da0d6-d597-4318-ba7f-da9c4d32e518\") " Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.488405 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs" (OuterVolumeSpecName: "kube-api-access-46pjs") pod "e29da0d6-d597-4318-ba7f-da9c4d32e518" (UID: "e29da0d6-d597-4318-ba7f-da9c4d32e518"). InnerVolumeSpecName "kube-api-access-46pjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:20 crc kubenswrapper[5002]: I1014 08:07:20.585331 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46pjs\" (UniqueName: \"kubernetes.io/projected/e29da0d6-d597-4318-ba7f-da9c4d32e518-kube-api-access-46pjs\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:21 crc kubenswrapper[5002]: I1014 08:07:21.260829 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4tkcz" event={"ID":"aa02e7c2-896a-4750-968d-9a4543651928","Type":"ContainerStarted","Data":"ce85c3c47b4c2af18034f1cf83a1567bb7d189cc076da9d15d5f3ba00bd817b3"} Oct 14 08:07:21 crc kubenswrapper[5002]: I1014 08:07:21.260863 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b87gc" Oct 14 08:07:21 crc kubenswrapper[5002]: I1014 08:07:21.291541 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4tkcz" podStartSLOduration=5.825077155 podStartE2EDuration="10.291524353s" podCreationTimestamp="2025-10-14 08:07:11 +0000 UTC" firstStartedPulling="2025-10-14 08:07:15.803089608 +0000 UTC m=+968.784329060" lastFinishedPulling="2025-10-14 08:07:20.269536786 +0000 UTC m=+973.250776258" observedRunningTime="2025-10-14 08:07:21.284898206 +0000 UTC m=+974.266137678" watchObservedRunningTime="2025-10-14 08:07:21.291524353 +0000 UTC m=+974.272763805" Oct 14 08:07:23 crc kubenswrapper[5002]: I1014 08:07:23.295258 5002 generic.go:334] "Generic (PLEG): container finished" podID="55b80ca7-5604-4c84-96f5-1279ecf2abe7" containerID="8f78daae4f3b6f9db13c31b5682eee751191639b981f868ce72155de15da63ac" exitCode=0 Oct 14 08:07:23 crc kubenswrapper[5002]: I1014 08:07:23.295342 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z4thp" event={"ID":"55b80ca7-5604-4c84-96f5-1279ecf2abe7","Type":"ContainerDied","Data":"8f78daae4f3b6f9db13c31b5682eee751191639b981f868ce72155de15da63ac"} Oct 14 08:07:23 crc kubenswrapper[5002]: I1014 08:07:23.298033 5002 generic.go:334] "Generic (PLEG): container finished" podID="aa02e7c2-896a-4750-968d-9a4543651928" containerID="ce85c3c47b4c2af18034f1cf83a1567bb7d189cc076da9d15d5f3ba00bd817b3" exitCode=0 Oct 14 08:07:23 crc kubenswrapper[5002]: I1014 08:07:23.298069 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4tkcz" event={"ID":"aa02e7c2-896a-4750-968d-9a4543651928","Type":"ContainerDied","Data":"ce85c3c47b4c2af18034f1cf83a1567bb7d189cc076da9d15d5f3ba00bd817b3"} Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.718499 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.759211 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data\") pod \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.759305 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle\") pod \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.759512 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8qqv\" (UniqueName: \"kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv\") pod \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.759556 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data\") pod \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\" (UID: \"55b80ca7-5604-4c84-96f5-1279ecf2abe7\") " Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.766033 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv" (OuterVolumeSpecName: "kube-api-access-c8qqv") pod "55b80ca7-5604-4c84-96f5-1279ecf2abe7" (UID: "55b80ca7-5604-4c84-96f5-1279ecf2abe7"). InnerVolumeSpecName "kube-api-access-c8qqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.766086 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "55b80ca7-5604-4c84-96f5-1279ecf2abe7" (UID: "55b80ca7-5604-4c84-96f5-1279ecf2abe7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.798785 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55b80ca7-5604-4c84-96f5-1279ecf2abe7" (UID: "55b80ca7-5604-4c84-96f5-1279ecf2abe7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.825709 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data" (OuterVolumeSpecName: "config-data") pod "55b80ca7-5604-4c84-96f5-1279ecf2abe7" (UID: "55b80ca7-5604-4c84-96f5-1279ecf2abe7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.867864 5002 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.867910 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.867923 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b80ca7-5604-4c84-96f5-1279ecf2abe7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:24 crc kubenswrapper[5002]: I1014 08:07:24.867935 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8qqv\" (UniqueName: \"kubernetes.io/projected/55b80ca7-5604-4c84-96f5-1279ecf2abe7-kube-api-access-c8qqv\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.316325 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-z4thp" event={"ID":"55b80ca7-5604-4c84-96f5-1279ecf2abe7","Type":"ContainerDied","Data":"d96c9f2281e39628e5d1e17ecfacaefb7c8934a4bbf55b39ed5fe62f497df524"} Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.316370 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d96c9f2281e39628e5d1e17ecfacaefb7c8934a4bbf55b39ed5fe62f497df524" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.316819 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-z4thp" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.357335 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.476652 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data\") pod \"aa02e7c2-896a-4750-968d-9a4543651928\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.476818 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcdcw\" (UniqueName: \"kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw\") pod \"aa02e7c2-896a-4750-968d-9a4543651928\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.476973 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle\") pod \"aa02e7c2-896a-4750-968d-9a4543651928\" (UID: \"aa02e7c2-896a-4750-968d-9a4543651928\") " Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.479898 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw" (OuterVolumeSpecName: "kube-api-access-hcdcw") pod "aa02e7c2-896a-4750-968d-9a4543651928" (UID: "aa02e7c2-896a-4750-968d-9a4543651928"). InnerVolumeSpecName "kube-api-access-hcdcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.503959 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa02e7c2-896a-4750-968d-9a4543651928" (UID: "aa02e7c2-896a-4750-968d-9a4543651928"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.531810 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data" (OuterVolumeSpecName: "config-data") pod "aa02e7c2-896a-4750-968d-9a4543651928" (UID: "aa02e7c2-896a-4750-968d-9a4543651928"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.578419 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.578458 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa02e7c2-896a-4750-968d-9a4543651928-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.578471 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcdcw\" (UniqueName: \"kubernetes.io/projected/aa02e7c2-896a-4750-968d-9a4543651928-kube-api-access-hcdcw\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.695936 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:25 crc kubenswrapper[5002]: E1014 08:07:25.696353 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696375 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: E1014 08:07:25.696394 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b80ca7-5604-4c84-96f5-1279ecf2abe7" containerName="glance-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696402 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b80ca7-5604-4c84-96f5-1279ecf2abe7" containerName="glance-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: E1014 08:07:25.696422 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e29da0d6-d597-4318-ba7f-da9c4d32e518" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696430 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e29da0d6-d597-4318-ba7f-da9c4d32e518" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: E1014 08:07:25.696455 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa02e7c2-896a-4750-968d-9a4543651928" containerName="keystone-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696464 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa02e7c2-896a-4750-968d-9a4543651928" containerName="keystone-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: E1014 08:07:25.696479 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3cb8568-04da-484f-a4f8-f1b2e8ab755c" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696486 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3cb8568-04da-484f-a4f8-f1b2e8ab755c" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696685 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b80ca7-5604-4c84-96f5-1279ecf2abe7" containerName="glance-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696703 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3cb8568-04da-484f-a4f8-f1b2e8ab755c" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696712 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696730 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e29da0d6-d597-4318-ba7f-da9c4d32e518" containerName="mariadb-database-create" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.696747 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa02e7c2-896a-4750-968d-9a4543651928" containerName="keystone-db-sync" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.697688 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.706613 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.782213 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.782253 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.782285 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.782321 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.782373 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk787\" (UniqueName: \"kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.883929 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.884299 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.884387 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk787\" (UniqueName: \"kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.884455 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.884485 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.885261 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.885427 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.885676 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.885700 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:25 crc kubenswrapper[5002]: I1014 08:07:25.902072 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk787\" (UniqueName: \"kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787\") pod \"dnsmasq-dns-54c4767797-mqq4z\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.016013 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.325775 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4tkcz" event={"ID":"aa02e7c2-896a-4750-968d-9a4543651928","Type":"ContainerDied","Data":"b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f"} Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.325823 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2a94a6123b27dbb0bce1e35853bfb57240a77a77888b8cd89f14414b3cf085f" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.325884 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4tkcz" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.462343 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.608405 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.631657 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.633218 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.651993 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nqkql"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.656036 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.658940 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.659175 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.659365 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jps45" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.660190 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.676334 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.677339 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nqkql"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.694917 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.695124 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxbp\" (UniqueName: \"kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.695261 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.695357 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.695504 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.799217 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.802523 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.811692 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.811987 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmtsm\" (UniqueName: \"kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812043 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxbp\" (UniqueName: \"kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812074 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812119 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812150 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812186 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812264 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812283 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812322 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.812401 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.813534 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.814263 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.814421 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.814622 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.814767 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.814909 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-lz8rd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.815072 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.815205 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.819749 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.848991 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxbp\" (UniqueName: \"kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp\") pod \"dnsmasq-dns-76f4594469-blktd\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.874888 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.876886 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.880421 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.880580 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.912800 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913653 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913690 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913721 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913768 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrlwb\" (UniqueName: \"kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913826 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913901 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913922 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmtsm\" (UniqueName: \"kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913952 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913977 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.913994 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.914021 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.917004 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.922445 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.936520 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.939170 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.943705 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.957210 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.991663 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmtsm\" (UniqueName: \"kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm\") pod \"keystone-bootstrap-nqkql\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.991931 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:07:26 crc kubenswrapper[5002]: I1014 08:07:26.993130 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.010276 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021564 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021609 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021639 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021667 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021686 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021714 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021739 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021758 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021783 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021819 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrlwb\" (UniqueName: \"kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021861 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.021891 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppgq8\" (UniqueName: \"kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.025166 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.025581 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.025781 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.026684 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.037338 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.058045 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrlwb\" (UniqueName: \"kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb\") pod \"horizon-577688b66f-svbn9\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.103713 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130090 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130157 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppgq8\" (UniqueName: \"kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130233 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130281 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhjff\" (UniqueName: \"kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130314 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130338 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130361 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130427 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130493 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130524 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130564 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.130624 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.132421 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.132620 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.140220 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.140865 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.150467 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.150491 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.151562 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.159717 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppgq8\" (UniqueName: \"kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8\") pod \"ceilometer-0\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.174473 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.175816 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.194417 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jv98l"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.200678 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.205908 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.206042 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.208046 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gvlg9" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.213744 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.214116 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233668 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233750 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233805 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233855 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwld7\" (UniqueName: \"kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233905 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.233937 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhjff\" (UniqueName: \"kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.234248 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.234270 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.234321 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.234397 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.235121 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.235352 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.236860 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jv98l"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.239023 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.239391 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.262478 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhjff\" (UniqueName: \"kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff\") pod \"horizon-d5dc898f-nqhn4\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.336761 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.336872 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwld7\" (UniqueName: \"kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.336900 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337277 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337349 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337372 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337493 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337570 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337614 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk7j6\" (UniqueName: \"kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337632 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.337730 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.338422 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.339072 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.339495 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.389148 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwld7\" (UniqueName: \"kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7\") pod \"dnsmasq-dns-6bd85975c7-6cwfb\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.398213 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.435181 5002 generic.go:334] "Generic (PLEG): container finished" podID="2fc3ddb8-a431-4255-9629-b3018683c964" containerID="1901f76d99956bb7c051b49dfa83fb70eb41e0ca9601b1fc14619271bf64ecd4" exitCode=0 Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.435219 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" event={"ID":"2fc3ddb8-a431-4255-9629-b3018683c964","Type":"ContainerDied","Data":"1901f76d99956bb7c051b49dfa83fb70eb41e0ca9601b1fc14619271bf64ecd4"} Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.435246 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" event={"ID":"2fc3ddb8-a431-4255-9629-b3018683c964","Type":"ContainerStarted","Data":"cd15c49c864a42a74c1f5ea818cd65e0566d4e86f85e1c7d6607a3a29e9a4610"} Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.441928 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk7j6\" (UniqueName: \"kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.442087 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.442106 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.442128 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.442148 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.443149 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.445561 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.448658 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.449539 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.481512 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk7j6\" (UniqueName: \"kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6\") pod \"placement-db-sync-jv98l\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.526190 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.528686 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nqkql"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.539040 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jv98l" Oct 14 08:07:27 crc kubenswrapper[5002]: W1014 08:07:27.557389 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod415b7b71_79f0_434a_8439_78c75f7a0dd2.slice/crio-5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db WatchSource:0}: Error finding container 5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db: Status 404 returned error can't find the container with id 5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.591293 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:27 crc kubenswrapper[5002]: W1014 08:07:27.595924 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod144040fc_e8aa_4d2c_a7c4_9cbaeb6ce42d.slice/crio-0a4765f40fa4fe61b7a0d51b2fb74dca4076f2ab28a86beb8262d2d9339f8d4b WatchSource:0}: Error finding container 0a4765f40fa4fe61b7a0d51b2fb74dca4076f2ab28a86beb8262d2d9339f8d4b: Status 404 returned error can't find the container with id 0a4765f40fa4fe61b7a0d51b2fb74dca4076f2ab28a86beb8262d2d9339f8d4b Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.804499 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.920965 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.930410 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:07:27 crc kubenswrapper[5002]: W1014 08:07:27.968193 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12720d49_a4e4_426e_bf38_169a0beeb8d2.slice/crio-df5f6319a166b0c7cf9152000b3df0631123ead38811d24faffc9654f437c6e6 WatchSource:0}: Error finding container df5f6319a166b0c7cf9152000b3df0631123ead38811d24faffc9654f437c6e6: Status 404 returned error can't find the container with id df5f6319a166b0c7cf9152000b3df0631123ead38811d24faffc9654f437c6e6 Oct 14 08:07:27 crc kubenswrapper[5002]: I1014 08:07:27.969669 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.007549 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jv98l"] Oct 14 08:07:28 crc kubenswrapper[5002]: W1014 08:07:28.010204 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfce7a28b_9ac5_4a32_a9c9_859df10c79c0.slice/crio-1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff WatchSource:0}: Error finding container 1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff: Status 404 returned error can't find the container with id 1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.055036 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config\") pod \"2fc3ddb8-a431-4255-9629-b3018683c964\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.055200 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb\") pod \"2fc3ddb8-a431-4255-9629-b3018683c964\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.055243 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc\") pod \"2fc3ddb8-a431-4255-9629-b3018683c964\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.055301 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb\") pod \"2fc3ddb8-a431-4255-9629-b3018683c964\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.055334 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk787\" (UniqueName: \"kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787\") pod \"2fc3ddb8-a431-4255-9629-b3018683c964\" (UID: \"2fc3ddb8-a431-4255-9629-b3018683c964\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.061035 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787" (OuterVolumeSpecName: "kube-api-access-nk787") pod "2fc3ddb8-a431-4255-9629-b3018683c964" (UID: "2fc3ddb8-a431-4255-9629-b3018683c964"). InnerVolumeSpecName "kube-api-access-nk787". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.078254 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2fc3ddb8-a431-4255-9629-b3018683c964" (UID: "2fc3ddb8-a431-4255-9629-b3018683c964"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.081299 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config" (OuterVolumeSpecName: "config") pod "2fc3ddb8-a431-4255-9629-b3018683c964" (UID: "2fc3ddb8-a431-4255-9629-b3018683c964"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.085231 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fc3ddb8-a431-4255-9629-b3018683c964" (UID: "2fc3ddb8-a431-4255-9629-b3018683c964"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.092323 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fc3ddb8-a431-4255-9629-b3018683c964" (UID: "2fc3ddb8-a431-4255-9629-b3018683c964"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.106355 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:07:28 crc kubenswrapper[5002]: W1014 08:07:28.110934 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8a10d6e_f132_43fe_ac06_cf1a99c9146f.slice/crio-677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50 WatchSource:0}: Error finding container 677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50: Status 404 returned error can't find the container with id 677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50 Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.158500 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.158541 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.158554 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.158565 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk787\" (UniqueName: \"kubernetes.io/projected/2fc3ddb8-a431-4255-9629-b3018683c964-kube-api-access-nk787\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.158582 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc3ddb8-a431-4255-9629-b3018683c964-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.451686 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" event={"ID":"2fc3ddb8-a431-4255-9629-b3018683c964","Type":"ContainerDied","Data":"cd15c49c864a42a74c1f5ea818cd65e0566d4e86f85e1c7d6607a3a29e9a4610"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.452029 5002 scope.go:117] "RemoveContainer" containerID="1901f76d99956bb7c051b49dfa83fb70eb41e0ca9601b1fc14619271bf64ecd4" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.451728 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54c4767797-mqq4z" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.453459 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerStarted","Data":"49aa15a0ced23cd81fd7acc3791150335541d212a15ea2475be209ec18843b56"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.455999 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqkql" event={"ID":"415b7b71-79f0-434a-8439-78c75f7a0dd2","Type":"ContainerStarted","Data":"ac854a13512a9d841b051aa8b108fe2f0839603d57044323c1936182255189ec"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.456033 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqkql" event={"ID":"415b7b71-79f0-434a-8439-78c75f7a0dd2","Type":"ContainerStarted","Data":"5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.458415 5002 generic.go:334] "Generic (PLEG): container finished" podID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerID="d928716ed2af799033fe6b4c73237e741f2698d058fb0a12896fb0628a0ddd20" exitCode=0 Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.458481 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" event={"ID":"e8a10d6e-f132-43fe-ac06-cf1a99c9146f","Type":"ContainerDied","Data":"d928716ed2af799033fe6b4c73237e741f2698d058fb0a12896fb0628a0ddd20"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.458507 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" event={"ID":"e8a10d6e-f132-43fe-ac06-cf1a99c9146f","Type":"ContainerStarted","Data":"677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.463781 5002 generic.go:334] "Generic (PLEG): container finished" podID="144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" containerID="5be74fed257047bec494e550d27ed2654eff7af99038a20d4b10d623d36a4271" exitCode=0 Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.463936 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76f4594469-blktd" event={"ID":"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d","Type":"ContainerDied","Data":"5be74fed257047bec494e550d27ed2654eff7af99038a20d4b10d623d36a4271"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.464009 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76f4594469-blktd" event={"ID":"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d","Type":"ContainerStarted","Data":"0a4765f40fa4fe61b7a0d51b2fb74dca4076f2ab28a86beb8262d2d9339f8d4b"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.469074 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jv98l" event={"ID":"fce7a28b-9ac5-4a32-a9c9-859df10c79c0","Type":"ContainerStarted","Data":"1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.476757 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerStarted","Data":"df5f6319a166b0c7cf9152000b3df0631123ead38811d24faffc9654f437c6e6"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.478487 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerStarted","Data":"317afa8bd12836ac2bcc3588f6613edd32c232146b1f149de98c5938c7acde9c"} Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.487299 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nqkql" podStartSLOduration=2.487281844 podStartE2EDuration="2.487281844s" podCreationTimestamp="2025-10-14 08:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:28.482996439 +0000 UTC m=+981.464235901" watchObservedRunningTime="2025-10-14 08:07:28.487281844 +0000 UTC m=+981.468521306" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.583982 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.589806 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54c4767797-mqq4z"] Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.836427 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.976691 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config\") pod \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.976744 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc\") pod \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.976770 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb\") pod \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.976995 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfxbp\" (UniqueName: \"kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp\") pod \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.977046 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb\") pod \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\" (UID: \"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d\") " Oct 14 08:07:28 crc kubenswrapper[5002]: I1014 08:07:28.983393 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp" (OuterVolumeSpecName: "kube-api-access-zfxbp") pod "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" (UID: "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d"). InnerVolumeSpecName "kube-api-access-zfxbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.015220 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" (UID: "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.018767 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config" (OuterVolumeSpecName: "config") pod "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" (UID: "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.024927 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" (UID: "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.043407 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" (UID: "144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.081385 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.082071 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.082188 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfxbp\" (UniqueName: \"kubernetes.io/projected/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-kube-api-access-zfxbp\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.082200 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.082747 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.490137 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76f4594469-blktd" event={"ID":"144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d","Type":"ContainerDied","Data":"0a4765f40fa4fe61b7a0d51b2fb74dca4076f2ab28a86beb8262d2d9339f8d4b"} Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.490416 5002 scope.go:117] "RemoveContainer" containerID="5be74fed257047bec494e550d27ed2654eff7af99038a20d4b10d623d36a4271" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.490173 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76f4594469-blktd" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.496403 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" event={"ID":"e8a10d6e-f132-43fe-ac06-cf1a99c9146f","Type":"ContainerStarted","Data":"46badaaa8912c6af3248be56592226d71d3ebc94ebd469fac2d66d7e695c6261"} Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.496462 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.511303 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" podStartSLOduration=2.511284277 podStartE2EDuration="2.511284277s" podCreationTimestamp="2025-10-14 08:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:29.510345391 +0000 UTC m=+982.491584863" watchObservedRunningTime="2025-10-14 08:07:29.511284277 +0000 UTC m=+982.492523729" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.547354 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.552355 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76f4594469-blktd"] Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.734358 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" path="/var/lib/kubelet/pods/144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d/volumes" Oct 14 08:07:29 crc kubenswrapper[5002]: I1014 08:07:29.734987 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc3ddb8-a431-4255-9629-b3018683c964" path="/var/lib/kubelet/pods/2fc3ddb8-a431-4255-9629-b3018683c964/volumes" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.057245 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.062980 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.103606 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:07:30 crc kubenswrapper[5002]: E1014 08:07:30.103947 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc3ddb8-a431-4255-9629-b3018683c964" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.103959 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc3ddb8-a431-4255-9629-b3018683c964" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: E1014 08:07:30.103971 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.103977 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.104140 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc3ddb8-a431-4255-9629-b3018683c964" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.104162 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="144040fc-e8aa-4d2c-a7c4-9cbaeb6ce42d" containerName="init" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.105009 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.115516 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.202117 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.202197 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.202217 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.202486 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x8s5\" (UniqueName: \"kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.202611 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.304024 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x8s5\" (UniqueName: \"kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.304083 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.304137 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.304221 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.304242 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.305094 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.305413 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.305757 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.316458 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.320948 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x8s5\" (UniqueName: \"kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5\") pod \"horizon-77cd4c5db5-24j69\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:30 crc kubenswrapper[5002]: I1014 08:07:30.421331 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.154398 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-78b8-account-create-cx2pn"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.155561 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.157255 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.160904 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-78b8-account-create-cx2pn"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.231754 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6clz4\" (UniqueName: \"kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4\") pod \"cinder-78b8-account-create-cx2pn\" (UID: \"3b0300de-4b38-4615-afd4-910e29e28127\") " pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.258779 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-56ab-account-create-64pzr"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.260461 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.263054 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.272433 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-56ab-account-create-64pzr"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.333109 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j76f7\" (UniqueName: \"kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7\") pod \"barbican-56ab-account-create-64pzr\" (UID: \"84d1d7d9-db0a-47a7-b8de-23b98786f876\") " pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.333168 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6clz4\" (UniqueName: \"kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4\") pod \"cinder-78b8-account-create-cx2pn\" (UID: \"3b0300de-4b38-4615-afd4-910e29e28127\") " pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.354006 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6clz4\" (UniqueName: \"kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4\") pod \"cinder-78b8-account-create-cx2pn\" (UID: \"3b0300de-4b38-4615-afd4-910e29e28127\") " pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.442210 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j76f7\" (UniqueName: \"kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7\") pod \"barbican-56ab-account-create-64pzr\" (UID: \"84d1d7d9-db0a-47a7-b8de-23b98786f876\") " pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.449552 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f03f-account-create-6kv52"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.450772 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.452749 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.457051 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f03f-account-create-6kv52"] Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.466574 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j76f7\" (UniqueName: \"kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7\") pod \"barbican-56ab-account-create-64pzr\" (UID: \"84d1d7d9-db0a-47a7-b8de-23b98786f876\") " pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.530419 5002 generic.go:334] "Generic (PLEG): container finished" podID="415b7b71-79f0-434a-8439-78c75f7a0dd2" containerID="ac854a13512a9d841b051aa8b108fe2f0839603d57044323c1936182255189ec" exitCode=0 Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.531281 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqkql" event={"ID":"415b7b71-79f0-434a-8439-78c75f7a0dd2","Type":"ContainerDied","Data":"ac854a13512a9d841b051aa8b108fe2f0839603d57044323c1936182255189ec"} Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.535902 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.550052 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntt54\" (UniqueName: \"kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54\") pod \"neutron-f03f-account-create-6kv52\" (UID: \"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a\") " pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.575619 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.652090 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntt54\" (UniqueName: \"kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54\") pod \"neutron-f03f-account-create-6kv52\" (UID: \"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a\") " pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.673433 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntt54\" (UniqueName: \"kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54\") pod \"neutron-f03f-account-create-6kv52\" (UID: \"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a\") " pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:31 crc kubenswrapper[5002]: I1014 08:07:31.810665 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:37 crc kubenswrapper[5002]: I1014 08:07:37.527929 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:07:37 crc kubenswrapper[5002]: I1014 08:07:37.624762 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:07:37 crc kubenswrapper[5002]: I1014 08:07:37.625303 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="dnsmasq-dns" containerID="cri-o://ab30045065cc39c7fcf0a9122420726490537441cdae94dd465a50fe3e9cd320" gracePeriod=10 Oct 14 08:07:38 crc kubenswrapper[5002]: I1014 08:07:38.628201 5002 generic.go:334] "Generic (PLEG): container finished" podID="f836312e-c43a-47ed-bfd8-a464948d9687" containerID="ab30045065cc39c7fcf0a9122420726490537441cdae94dd465a50fe3e9cd320" exitCode=0 Oct 14 08:07:38 crc kubenswrapper[5002]: I1014 08:07:38.628301 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" event={"ID":"f836312e-c43a-47ed-bfd8-a464948d9687","Type":"ContainerDied","Data":"ab30045065cc39c7fcf0a9122420726490537441cdae94dd465a50fe3e9cd320"} Oct 14 08:07:38 crc kubenswrapper[5002]: I1014 08:07:38.810753 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.538199 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.572777 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.574079 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.577233 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.583125 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.626094 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.653175 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-557d55dc9d-sk8l2"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.668189 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-557d55dc9d-sk8l2"] Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.668317 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710031 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710092 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710166 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710195 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710219 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710242 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44nqs\" (UniqueName: \"kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.710260 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mnv5\" (UniqueName: \"kubernetes.io/projected/681029aa-9883-4641-9008-aeb673358cdd-kube-api-access-9mnv5\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812605 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812667 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812704 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812735 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44nqs\" (UniqueName: \"kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812756 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812791 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-secret-key\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812891 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812941 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.812978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-tls-certs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.813033 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/681029aa-9883-4641-9008-aeb673358cdd-logs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.813061 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-config-data\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.813084 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-scripts\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.813106 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-combined-ca-bundle\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.814244 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.815758 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.815827 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.823107 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.823359 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.824164 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.837818 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44nqs\" (UniqueName: \"kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs\") pod \"horizon-5b7cd67f54-t9dgw\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.901419 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914291 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-secret-key\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914388 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-tls-certs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914442 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/681029aa-9883-4641-9008-aeb673358cdd-logs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914468 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-config-data\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914493 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-scripts\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914516 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-combined-ca-bundle\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914557 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mnv5\" (UniqueName: \"kubernetes.io/projected/681029aa-9883-4641-9008-aeb673358cdd-kube-api-access-9mnv5\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.914926 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/681029aa-9883-4641-9008-aeb673358cdd-logs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.915270 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-scripts\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.915930 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/681029aa-9883-4641-9008-aeb673358cdd-config-data\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.918263 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-secret-key\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.921935 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-horizon-tls-certs\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.922002 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/681029aa-9883-4641-9008-aeb673358cdd-combined-ca-bundle\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.934649 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mnv5\" (UniqueName: \"kubernetes.io/projected/681029aa-9883-4641-9008-aeb673358cdd-kube-api-access-9mnv5\") pod \"horizon-557d55dc9d-sk8l2\" (UID: \"681029aa-9883-4641-9008-aeb673358cdd\") " pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:39 crc kubenswrapper[5002]: I1014 08:07:39.985682 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:40 crc kubenswrapper[5002]: E1014 08:07:40.149375 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c" Oct 14 08:07:40 crc kubenswrapper[5002]: E1014 08:07:40.149576 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sk7j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-jv98l_openstack(fce7a28b-9ac5-4a32-a9c9-859df10c79c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:07:40 crc kubenswrapper[5002]: E1014 08:07:40.151012 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-jv98l" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.237007 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322129 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322430 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmtsm\" (UniqueName: \"kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322513 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322556 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322636 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.322669 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys\") pod \"415b7b71-79f0-434a-8439-78c75f7a0dd2\" (UID: \"415b7b71-79f0-434a-8439-78c75f7a0dd2\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.325463 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.326915 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts" (OuterVolumeSpecName: "scripts") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.332324 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm" (OuterVolumeSpecName: "kube-api-access-kmtsm") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "kube-api-access-kmtsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.332595 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.361011 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.367077 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data" (OuterVolumeSpecName: "config-data") pod "415b7b71-79f0-434a-8439-78c75f7a0dd2" (UID: "415b7b71-79f0-434a-8439-78c75f7a0dd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425034 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425066 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425081 5002 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425091 5002 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425101 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/415b7b71-79f0-434a-8439-78c75f7a0dd2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.425112 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmtsm\" (UniqueName: \"kubernetes.io/projected/415b7b71-79f0-434a-8439-78c75f7a0dd2-kube-api-access-kmtsm\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.565326 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.681440 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerStarted","Data":"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550"} Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.697763 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" event={"ID":"f836312e-c43a-47ed-bfd8-a464948d9687","Type":"ContainerDied","Data":"7095736e637377fbd24a00ba797fcb5e091b3f43082a8fddc267369579980dd2"} Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.697819 5002 scope.go:117] "RemoveContainer" containerID="ab30045065cc39c7fcf0a9122420726490537441cdae94dd465a50fe3e9cd320" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.697774 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-bw7l6" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.699137 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerStarted","Data":"542d9dc966b99803bafeed52ab3f95e09b26fa4621c3c7ba2b413b0a920b7ab9"} Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.706960 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nqkql" event={"ID":"415b7b71-79f0-434a-8439-78c75f7a0dd2","Type":"ContainerDied","Data":"5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db"} Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.706999 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fcea8afdef4760bc4f4e0396490255d8b9b9709509209526bd025a5f62794db" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.707074 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nqkql" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.713658 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerStarted","Data":"6713e666d17a7b8c0557c873580a43ff3fd96b008965abade3601b40a1398a83"} Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.726216 5002 scope.go:117] "RemoveContainer" containerID="46fe38a4f5bd2fc6926403544eba335c1806269d42a56a01e2fd5c41aab1aa07" Oct 14 08:07:40 crc kubenswrapper[5002]: E1014 08:07:40.726322 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c\\\"\"" pod="openstack/placement-db-sync-jv98l" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.728732 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb\") pod \"f836312e-c43a-47ed-bfd8-a464948d9687\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.728780 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc\") pod \"f836312e-c43a-47ed-bfd8-a464948d9687\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.728915 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb\") pod \"f836312e-c43a-47ed-bfd8-a464948d9687\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.728945 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config\") pod \"f836312e-c43a-47ed-bfd8-a464948d9687\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.729011 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gsb5\" (UniqueName: \"kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5\") pod \"f836312e-c43a-47ed-bfd8-a464948d9687\" (UID: \"f836312e-c43a-47ed-bfd8-a464948d9687\") " Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.735957 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5" (OuterVolumeSpecName: "kube-api-access-8gsb5") pod "f836312e-c43a-47ed-bfd8-a464948d9687" (UID: "f836312e-c43a-47ed-bfd8-a464948d9687"). InnerVolumeSpecName "kube-api-access-8gsb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.831020 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gsb5\" (UniqueName: \"kubernetes.io/projected/f836312e-c43a-47ed-bfd8-a464948d9687-kube-api-access-8gsb5\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.864398 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config" (OuterVolumeSpecName: "config") pod "f836312e-c43a-47ed-bfd8-a464948d9687" (UID: "f836312e-c43a-47ed-bfd8-a464948d9687"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.870136 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f03f-account-create-6kv52"] Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.873922 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f836312e-c43a-47ed-bfd8-a464948d9687" (UID: "f836312e-c43a-47ed-bfd8-a464948d9687"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.878179 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f836312e-c43a-47ed-bfd8-a464948d9687" (UID: "f836312e-c43a-47ed-bfd8-a464948d9687"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:40 crc kubenswrapper[5002]: I1014 08:07:40.879509 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f836312e-c43a-47ed-bfd8-a464948d9687" (UID: "f836312e-c43a-47ed-bfd8-a464948d9687"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.934319 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.934639 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.934649 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.934658 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836312e-c43a-47ed-bfd8-a464948d9687-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.938070 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-78b8-account-create-cx2pn"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.946444 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-56ab-account-create-64pzr"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:40.955818 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.081528 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-557d55dc9d-sk8l2"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.096997 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:07:41 crc kubenswrapper[5002]: W1014 08:07:41.102480 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e99bd9e_3960_439c_a053_f0ecdde86ccb.slice/crio-671ba1175ccc5bb8f6779cc658eb5f5dfb6bd7c74f2e941581393bd8e545f2fe WatchSource:0}: Error finding container 671ba1175ccc5bb8f6779cc658eb5f5dfb6bd7c74f2e941581393bd8e545f2fe: Status 404 returned error can't find the container with id 671ba1175ccc5bb8f6779cc658eb5f5dfb6bd7c74f2e941581393bd8e545f2fe Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.113937 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.119809 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-bw7l6"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.326357 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nqkql"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.335058 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nqkql"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.414867 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-snwzv"] Oct 14 08:07:41 crc kubenswrapper[5002]: E1014 08:07:41.415198 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="dnsmasq-dns" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415209 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="dnsmasq-dns" Oct 14 08:07:41 crc kubenswrapper[5002]: E1014 08:07:41.415224 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415b7b71-79f0-434a-8439-78c75f7a0dd2" containerName="keystone-bootstrap" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415231 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="415b7b71-79f0-434a-8439-78c75f7a0dd2" containerName="keystone-bootstrap" Oct 14 08:07:41 crc kubenswrapper[5002]: E1014 08:07:41.415240 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="init" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415245 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="init" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415387 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" containerName="dnsmasq-dns" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415413 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="415b7b71-79f0-434a-8439-78c75f7a0dd2" containerName="keystone-bootstrap" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.415960 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.423320 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.423761 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.425507 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-snwzv"] Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.427117 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jps45" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.427701 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546429 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546538 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546577 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546608 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mfmr\" (UniqueName: \"kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546641 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.546703 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648249 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648294 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648322 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mfmr\" (UniqueName: \"kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648344 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648366 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.648450 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.657759 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.660672 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.660143 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.663488 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.668647 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.673418 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mfmr\" (UniqueName: \"kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr\") pod \"keystone-bootstrap-snwzv\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.753214 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="415b7b71-79f0-434a-8439-78c75f7a0dd2" path="/var/lib/kubelet/pods/415b7b71-79f0-434a-8439-78c75f7a0dd2/volumes" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.753919 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f836312e-c43a-47ed-bfd8-a464948d9687" path="/var/lib/kubelet/pods/f836312e-c43a-47ed-bfd8-a464948d9687/volumes" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.757879 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-557d55dc9d-sk8l2" event={"ID":"681029aa-9883-4641-9008-aeb673358cdd","Type":"ContainerStarted","Data":"69ec938f855e42dfb8035242f4981cd979cc604a0ba5c85ec074a5206cccf41b"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.757942 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-557d55dc9d-sk8l2" event={"ID":"681029aa-9883-4641-9008-aeb673358cdd","Type":"ContainerStarted","Data":"3b880f92009cb34a162eee95df55ecc8fe9510c7894f2a525a5d2f4e85594613"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.766478 5002 generic.go:334] "Generic (PLEG): container finished" podID="ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" containerID="f9a847a10bd556ae01dc5e8adcc76f5b8bc558d27465de25cb82f9861a0c149e" exitCode=0 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.766574 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f03f-account-create-6kv52" event={"ID":"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a","Type":"ContainerDied","Data":"f9a847a10bd556ae01dc5e8adcc76f5b8bc558d27465de25cb82f9861a0c149e"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.766627 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f03f-account-create-6kv52" event={"ID":"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a","Type":"ContainerStarted","Data":"b6ef852f68b7ebbfd433bb0bcbd12115b9c2db9151de67173e1d28aae7e9dfef"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.772125 5002 generic.go:334] "Generic (PLEG): container finished" podID="84d1d7d9-db0a-47a7-b8de-23b98786f876" containerID="879e0281f9f2c43346dd40061861ddcbe7d0490d0d31eae37489dd2fd5132675" exitCode=0 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.772185 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-56ab-account-create-64pzr" event={"ID":"84d1d7d9-db0a-47a7-b8de-23b98786f876","Type":"ContainerDied","Data":"879e0281f9f2c43346dd40061861ddcbe7d0490d0d31eae37489dd2fd5132675"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.772208 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-56ab-account-create-64pzr" event={"ID":"84d1d7d9-db0a-47a7-b8de-23b98786f876","Type":"ContainerStarted","Data":"75abcf3c1979f0ca599382ea7d2e01b6f78d4c3de30b3d633b9b694dc18405da"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.780625 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerStarted","Data":"783cb420ade93fc94cb405135847f53d76475fa05bb6ec01bbe8cac8778c707e"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.782460 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d5dc898f-nqhn4" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon-log" containerID="cri-o://6713e666d17a7b8c0557c873580a43ff3fd96b008965abade3601b40a1398a83" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.782827 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d5dc898f-nqhn4" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon" containerID="cri-o://783cb420ade93fc94cb405135847f53d76475fa05bb6ec01bbe8cac8778c707e" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.789750 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerStarted","Data":"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.789793 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerStarted","Data":"968cf559e1c3aca3a2260b46db5cf82aaab1cc5e2cd652c788e13575bf224f56"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.800822 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerStarted","Data":"1717340cfb5720d63c8157ee059c297fe8ebbf217ddc80577f010d1a291ec26b"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.800968 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-577688b66f-svbn9" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon-log" containerID="cri-o://542d9dc966b99803bafeed52ab3f95e09b26fa4621c3c7ba2b413b0a920b7ab9" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.801237 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-577688b66f-svbn9" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon" containerID="cri-o://1717340cfb5720d63c8157ee059c297fe8ebbf217ddc80577f010d1a291ec26b" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.803074 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerStarted","Data":"f303ff5fb8c6acb8296f9ead1a3e6a64ea6507b6cfbb4b864a1d46b3de3bd8db"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.803106 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerStarted","Data":"671ba1175ccc5bb8f6779cc658eb5f5dfb6bd7c74f2e941581393bd8e545f2fe"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.803179 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77cd4c5db5-24j69" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon-log" containerID="cri-o://f303ff5fb8c6acb8296f9ead1a3e6a64ea6507b6cfbb4b864a1d46b3de3bd8db" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.803230 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77cd4c5db5-24j69" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon" containerID="cri-o://995104a78065f95a3691c5e595209bf467b2d52e6a7ae0e3ccabca98bd6178e9" gracePeriod=30 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.806625 5002 generic.go:334] "Generic (PLEG): container finished" podID="3b0300de-4b38-4615-afd4-910e29e28127" containerID="7c66370491ebe1acecc02c82a0d9aa51c3410f1027234cc6163a324a8544eca9" exitCode=0 Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.806701 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-78b8-account-create-cx2pn" event={"ID":"3b0300de-4b38-4615-afd4-910e29e28127","Type":"ContainerDied","Data":"7c66370491ebe1acecc02c82a0d9aa51c3410f1027234cc6163a324a8544eca9"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.806722 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-78b8-account-create-cx2pn" event={"ID":"3b0300de-4b38-4615-afd4-910e29e28127","Type":"ContainerStarted","Data":"d52afc2b31b53f3d8428ebec914f75507b5a73ebb54974b205b9b0c0be6b4bd2"} Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.816388 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5b7cd67f54-t9dgw" podStartSLOduration=2.816371953 podStartE2EDuration="2.816371953s" podCreationTimestamp="2025-10-14 08:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:41.814329789 +0000 UTC m=+994.795569261" watchObservedRunningTime="2025-10-14 08:07:41.816371953 +0000 UTC m=+994.797611395" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.852323 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d5dc898f-nqhn4" podStartSLOduration=3.5212682539999998 podStartE2EDuration="15.852290627s" podCreationTimestamp="2025-10-14 08:07:26 +0000 UTC" firstStartedPulling="2025-10-14 08:07:27.973358181 +0000 UTC m=+980.954597673" lastFinishedPulling="2025-10-14 08:07:40.304380574 +0000 UTC m=+993.285620046" observedRunningTime="2025-10-14 08:07:41.845502675 +0000 UTC m=+994.826742137" watchObservedRunningTime="2025-10-14 08:07:41.852290627 +0000 UTC m=+994.833530079" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.877901 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-577688b66f-svbn9" podStartSLOduration=3.483742328 podStartE2EDuration="15.877878063s" podCreationTimestamp="2025-10-14 08:07:26 +0000 UTC" firstStartedPulling="2025-10-14 08:07:27.956762156 +0000 UTC m=+980.938001608" lastFinishedPulling="2025-10-14 08:07:40.350897891 +0000 UTC m=+993.332137343" observedRunningTime="2025-10-14 08:07:41.86846541 +0000 UTC m=+994.849704872" watchObservedRunningTime="2025-10-14 08:07:41.877878063 +0000 UTC m=+994.859117525" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.903213 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77cd4c5db5-24j69" podStartSLOduration=11.903189962 podStartE2EDuration="11.903189962s" podCreationTimestamp="2025-10-14 08:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:41.886662718 +0000 UTC m=+994.867902190" watchObservedRunningTime="2025-10-14 08:07:41.903189962 +0000 UTC m=+994.884429424" Oct 14 08:07:41 crc kubenswrapper[5002]: I1014 08:07:41.955049 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.764207 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-snwzv"] Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.823652 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-snwzv" event={"ID":"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f","Type":"ContainerStarted","Data":"26ccd6b4faa1bfafab17db2cfb98477cf8c519464ffcea9c60c0f655081e0f00"} Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.825395 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerStarted","Data":"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2"} Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.827660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerStarted","Data":"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f"} Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.829164 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-557d55dc9d-sk8l2" event={"ID":"681029aa-9883-4641-9008-aeb673358cdd","Type":"ContainerStarted","Data":"bfffb5dacb7fe8955a55cc7f27a8d5c9e375bcc22005011a8e634144e8074768"} Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.832651 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerStarted","Data":"995104a78065f95a3691c5e595209bf467b2d52e6a7ae0e3ccabca98bd6178e9"} Oct 14 08:07:42 crc kubenswrapper[5002]: I1014 08:07:42.858479 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-557d55dc9d-sk8l2" podStartSLOduration=3.858458041 podStartE2EDuration="3.858458041s" podCreationTimestamp="2025-10-14 08:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:42.855757428 +0000 UTC m=+995.836996920" watchObservedRunningTime="2025-10-14 08:07:42.858458041 +0000 UTC m=+995.839697503" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.200761 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.280436 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntt54\" (UniqueName: \"kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54\") pod \"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a\" (UID: \"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a\") " Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.291003 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54" (OuterVolumeSpecName: "kube-api-access-ntt54") pod "ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" (UID: "ca8d232a-e9c2-404f-a09d-4c8e609c9b6a"). InnerVolumeSpecName "kube-api-access-ntt54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.382006 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntt54\" (UniqueName: \"kubernetes.io/projected/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a-kube-api-access-ntt54\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.382189 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.403007 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.483377 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j76f7\" (UniqueName: \"kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7\") pod \"84d1d7d9-db0a-47a7-b8de-23b98786f876\" (UID: \"84d1d7d9-db0a-47a7-b8de-23b98786f876\") " Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.483504 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6clz4\" (UniqueName: \"kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4\") pod \"3b0300de-4b38-4615-afd4-910e29e28127\" (UID: \"3b0300de-4b38-4615-afd4-910e29e28127\") " Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.488198 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7" (OuterVolumeSpecName: "kube-api-access-j76f7") pod "84d1d7d9-db0a-47a7-b8de-23b98786f876" (UID: "84d1d7d9-db0a-47a7-b8de-23b98786f876"). InnerVolumeSpecName "kube-api-access-j76f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.488906 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4" (OuterVolumeSpecName: "kube-api-access-6clz4") pod "3b0300de-4b38-4615-afd4-910e29e28127" (UID: "3b0300de-4b38-4615-afd4-910e29e28127"). InnerVolumeSpecName "kube-api-access-6clz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.586242 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j76f7\" (UniqueName: \"kubernetes.io/projected/84d1d7d9-db0a-47a7-b8de-23b98786f876-kube-api-access-j76f7\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.586275 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6clz4\" (UniqueName: \"kubernetes.io/projected/3b0300de-4b38-4615-afd4-910e29e28127-kube-api-access-6clz4\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.844889 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-78b8-account-create-cx2pn" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.844829 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-78b8-account-create-cx2pn" event={"ID":"3b0300de-4b38-4615-afd4-910e29e28127","Type":"ContainerDied","Data":"d52afc2b31b53f3d8428ebec914f75507b5a73ebb54974b205b9b0c0be6b4bd2"} Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.845273 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d52afc2b31b53f3d8428ebec914f75507b5a73ebb54974b205b9b0c0be6b4bd2" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.852518 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f03f-account-create-6kv52" event={"ID":"ca8d232a-e9c2-404f-a09d-4c8e609c9b6a","Type":"ContainerDied","Data":"b6ef852f68b7ebbfd433bb0bcbd12115b9c2db9151de67173e1d28aae7e9dfef"} Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.852558 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6ef852f68b7ebbfd433bb0bcbd12115b9c2db9151de67173e1d28aae7e9dfef" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.852609 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f03f-account-create-6kv52" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.855919 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-56ab-account-create-64pzr" event={"ID":"84d1d7d9-db0a-47a7-b8de-23b98786f876","Type":"ContainerDied","Data":"75abcf3c1979f0ca599382ea7d2e01b6f78d4c3de30b3d633b9b694dc18405da"} Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.855942 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75abcf3c1979f0ca599382ea7d2e01b6f78d4c3de30b3d633b9b694dc18405da" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.855980 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-56ab-account-create-64pzr" Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.859727 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-snwzv" event={"ID":"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f","Type":"ContainerStarted","Data":"2c2aae10edcba2d9bb02e51e4ff61a66699822f5f8ac4830c3ae5669ddb1e5e6"} Oct 14 08:07:43 crc kubenswrapper[5002]: I1014 08:07:43.887983 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-snwzv" podStartSLOduration=2.887961742 podStartE2EDuration="2.887961742s" podCreationTimestamp="2025-10-14 08:07:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:43.887381855 +0000 UTC m=+996.868621317" watchObservedRunningTime="2025-10-14 08:07:43.887961742 +0000 UTC m=+996.869201204" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.475196 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-v746c"] Oct 14 08:07:46 crc kubenswrapper[5002]: E1014 08:07:46.475799 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d1d7d9-db0a-47a7-b8de-23b98786f876" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.475813 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d1d7d9-db0a-47a7-b8de-23b98786f876" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: E1014 08:07:46.475857 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b0300de-4b38-4615-afd4-910e29e28127" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.475863 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b0300de-4b38-4615-afd4-910e29e28127" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: E1014 08:07:46.475870 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.475876 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.476043 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d1d7d9-db0a-47a7-b8de-23b98786f876" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.476062 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.476077 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b0300de-4b38-4615-afd4-910e29e28127" containerName="mariadb-account-create" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.476584 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.478565 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xbc8s" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.479994 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.480477 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.491185 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-v746c"] Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537446 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537505 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtj4l\" (UniqueName: \"kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537671 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537721 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.537752 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639529 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639607 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639644 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639667 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639714 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.639733 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtj4l\" (UniqueName: \"kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.640396 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.646251 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.655379 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.655785 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.655798 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.663458 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtj4l\" (UniqueName: \"kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l\") pod \"cinder-db-sync-v746c\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.735151 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-fbcgt"] Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.736201 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.738002 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rtxxg" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.744728 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.745009 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fbcgt"] Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.797826 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-v746c" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.844661 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.844741 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.844848 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdqxb\" (UniqueName: \"kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.881286 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-7wzf7"] Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.882502 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.884592 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.884699 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.884916 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vlswn" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.893765 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-7wzf7"] Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946267 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdqxb\" (UniqueName: \"kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946316 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946355 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946372 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6xqn\" (UniqueName: \"kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946422 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.946498 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.952126 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.952281 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:46 crc kubenswrapper[5002]: I1014 08:07:46.969588 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdqxb\" (UniqueName: \"kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb\") pod \"barbican-db-sync-fbcgt\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.047909 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.048039 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.048065 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6xqn\" (UniqueName: \"kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.051260 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.054722 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.059312 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.068379 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6xqn\" (UniqueName: \"kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn\") pod \"neutron-db-sync-7wzf7\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.142119 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.196874 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:07:47 crc kubenswrapper[5002]: I1014 08:07:47.399512 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:07:49 crc kubenswrapper[5002]: I1014 08:07:49.902251 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:49 crc kubenswrapper[5002]: I1014 08:07:49.902636 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:07:49 crc kubenswrapper[5002]: I1014 08:07:49.986581 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:49 crc kubenswrapper[5002]: I1014 08:07:49.986674 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:07:50 crc kubenswrapper[5002]: I1014 08:07:50.422210 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:07:50 crc kubenswrapper[5002]: I1014 08:07:50.928029 5002 generic.go:334] "Generic (PLEG): container finished" podID="9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" containerID="2c2aae10edcba2d9bb02e51e4ff61a66699822f5f8ac4830c3ae5669ddb1e5e6" exitCode=0 Oct 14 08:07:50 crc kubenswrapper[5002]: I1014 08:07:50.928057 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-snwzv" event={"ID":"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f","Type":"ContainerDied","Data":"2c2aae10edcba2d9bb02e51e4ff61a66699822f5f8ac4830c3ae5669ddb1e5e6"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.351505 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:52 crc kubenswrapper[5002]: W1014 08:07:52.354045 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bbe80a8_92fa_4f55_9937_414e1755a05e.slice/crio-d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf WatchSource:0}: Error finding container d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf: Status 404 returned error can't find the container with id d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.355724 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-7wzf7"] Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.516541 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fbcgt"] Oct 14 08:07:52 crc kubenswrapper[5002]: W1014 08:07:52.517879 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5131d0b_1f4b_4882_8dc9_ffaaa3a18d69.slice/crio-31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da WatchSource:0}: Error finding container 31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da: Status 404 returned error can't find the container with id 31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.523448 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-v746c"] Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548440 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548488 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mfmr\" (UniqueName: \"kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548596 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548640 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548683 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.548724 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys\") pod \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\" (UID: \"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f\") " Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.553755 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.554268 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.555864 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr" (OuterVolumeSpecName: "kube-api-access-4mfmr") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "kube-api-access-4mfmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.556256 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts" (OuterVolumeSpecName: "scripts") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.572395 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.580004 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data" (OuterVolumeSpecName: "config-data") pod "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" (UID: "9c28ed22-e6c7-4a08-ba4f-eb8052296a5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650635 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650664 5002 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650672 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650681 5002 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650689 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.650697 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mfmr\" (UniqueName: \"kubernetes.io/projected/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f-kube-api-access-4mfmr\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.965314 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-v746c" event={"ID":"0307318a-577d-4b25-9972-274b46d62f36","Type":"ContainerStarted","Data":"79b412ee52c9513c5c3a15c3cb804a430281432665ad30341481316e6456ea88"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.966546 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7wzf7" event={"ID":"8bbe80a8-92fa-4f55-9937-414e1755a05e","Type":"ContainerStarted","Data":"77a7cca8e565256e664d13193a99ac15aafc2b59ed7c2f29a887c5018acc71cd"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.966572 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7wzf7" event={"ID":"8bbe80a8-92fa-4f55-9937-414e1755a05e","Type":"ContainerStarted","Data":"d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.975239 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fbcgt" event={"ID":"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69","Type":"ContainerStarted","Data":"31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.979654 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-snwzv" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.979664 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-snwzv" event={"ID":"9c28ed22-e6c7-4a08-ba4f-eb8052296a5f","Type":"ContainerDied","Data":"26ccd6b4faa1bfafab17db2cfb98477cf8c519464ffcea9c60c0f655081e0f00"} Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.979707 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26ccd6b4faa1bfafab17db2cfb98477cf8c519464ffcea9c60c0f655081e0f00" Oct 14 08:07:52 crc kubenswrapper[5002]: I1014 08:07:52.987782 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerStarted","Data":"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2"} Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.007508 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-7wzf7" podStartSLOduration=7.007486936 podStartE2EDuration="7.007486936s" podCreationTimestamp="2025-10-14 08:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:53.000219001 +0000 UTC m=+1005.981458463" watchObservedRunningTime="2025-10-14 08:07:53.007486936 +0000 UTC m=+1005.988726388" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.036110 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7db98f685f-fphzv"] Oct 14 08:07:53 crc kubenswrapper[5002]: E1014 08:07:53.036677 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" containerName="keystone-bootstrap" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.036698 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" containerName="keystone-bootstrap" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.037019 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" containerName="keystone-bootstrap" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.037927 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.049448 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.051151 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jps45" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.051179 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.051217 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.051386 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.051389 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.057467 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7db98f685f-fphzv"] Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.077970 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-public-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078028 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdxht\" (UniqueName: \"kubernetes.io/projected/e587d08e-64fa-4e35-98dc-ee283ada8c7f-kube-api-access-gdxht\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078094 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-config-data\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078124 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-credential-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078153 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-combined-ca-bundle\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078173 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-scripts\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078219 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-fernet-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.078248 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-internal-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179555 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-fernet-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179615 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-internal-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179661 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-public-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179680 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdxht\" (UniqueName: \"kubernetes.io/projected/e587d08e-64fa-4e35-98dc-ee283ada8c7f-kube-api-access-gdxht\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179738 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-config-data\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179767 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-credential-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179796 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-combined-ca-bundle\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.179813 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-scripts\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.185351 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-credential-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.185903 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-fernet-keys\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.185912 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-internal-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.186735 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-scripts\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.187700 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-public-tls-certs\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.188722 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-config-data\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.191354 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e587d08e-64fa-4e35-98dc-ee283ada8c7f-combined-ca-bundle\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.205397 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdxht\" (UniqueName: \"kubernetes.io/projected/e587d08e-64fa-4e35-98dc-ee283ada8c7f-kube-api-access-gdxht\") pod \"keystone-7db98f685f-fphzv\" (UID: \"e587d08e-64fa-4e35-98dc-ee283ada8c7f\") " pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.360316 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:53 crc kubenswrapper[5002]: I1014 08:07:53.858155 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7db98f685f-fphzv"] Oct 14 08:07:53 crc kubenswrapper[5002]: W1014 08:07:53.865352 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode587d08e_64fa_4e35_98dc_ee283ada8c7f.slice/crio-0e814dde32297c65febb9f3d69487ee8e786644b78ed1e22180670cbb6c1c5e5 WatchSource:0}: Error finding container 0e814dde32297c65febb9f3d69487ee8e786644b78ed1e22180670cbb6c1c5e5: Status 404 returned error can't find the container with id 0e814dde32297c65febb9f3d69487ee8e786644b78ed1e22180670cbb6c1c5e5 Oct 14 08:07:54 crc kubenswrapper[5002]: I1014 08:07:54.034012 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7db98f685f-fphzv" event={"ID":"e587d08e-64fa-4e35-98dc-ee283ada8c7f","Type":"ContainerStarted","Data":"0e814dde32297c65febb9f3d69487ee8e786644b78ed1e22180670cbb6c1c5e5"} Oct 14 08:07:55 crc kubenswrapper[5002]: I1014 08:07:55.043490 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7db98f685f-fphzv" event={"ID":"e587d08e-64fa-4e35-98dc-ee283ada8c7f","Type":"ContainerStarted","Data":"868cd255ad806362ef031226341ee138826a75e4ca57f6d7df093010e0425618"} Oct 14 08:07:55 crc kubenswrapper[5002]: I1014 08:07:55.043764 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:07:55 crc kubenswrapper[5002]: I1014 08:07:55.067678 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7db98f685f-fphzv" podStartSLOduration=2.067662376 podStartE2EDuration="2.067662376s" podCreationTimestamp="2025-10-14 08:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:07:55.062691174 +0000 UTC m=+1008.043930646" watchObservedRunningTime="2025-10-14 08:07:55.067662376 +0000 UTC m=+1008.048901828" Oct 14 08:07:59 crc kubenswrapper[5002]: I1014 08:07:59.904354 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Oct 14 08:07:59 crc kubenswrapper[5002]: I1014 08:07:59.990209 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-557d55dc9d-sk8l2" podUID="681029aa-9883-4641-9008-aeb673358cdd" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 14 08:08:07 crc kubenswrapper[5002]: I1014 08:08:07.152658 5002 generic.go:334] "Generic (PLEG): container finished" podID="8bbe80a8-92fa-4f55-9937-414e1755a05e" containerID="77a7cca8e565256e664d13193a99ac15aafc2b59ed7c2f29a887c5018acc71cd" exitCode=0 Oct 14 08:08:07 crc kubenswrapper[5002]: I1014 08:08:07.152787 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7wzf7" event={"ID":"8bbe80a8-92fa-4f55-9937-414e1755a05e","Type":"ContainerDied","Data":"77a7cca8e565256e664d13193a99ac15aafc2b59ed7c2f29a887c5018acc71cd"} Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.064183 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c" Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.065029 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sk7j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-jv98l_openstack(fce7a28b-9ac5-4a32-a9c9-859df10c79c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.066260 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-jv98l" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.086979 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48" Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.087231 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppgq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(7a5a0fa6-e889-4d56-b2c7-070aa5012a10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 08:08:12 crc kubenswrapper[5002]: E1014 08:08:12.089426 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.175506 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.215825 5002 generic.go:334] "Generic (PLEG): container finished" podID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerID="783cb420ade93fc94cb405135847f53d76475fa05bb6ec01bbe8cac8778c707e" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.215866 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerDied","Data":"783cb420ade93fc94cb405135847f53d76475fa05bb6ec01bbe8cac8778c707e"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.215906 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerDied","Data":"6713e666d17a7b8c0557c873580a43ff3fd96b008965abade3601b40a1398a83"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.215877 5002 generic.go:334] "Generic (PLEG): container finished" podID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerID="6713e666d17a7b8c0557c873580a43ff3fd96b008965abade3601b40a1398a83" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.219171 5002 generic.go:334] "Generic (PLEG): container finished" podID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerID="1717340cfb5720d63c8157ee059c297fe8ebbf217ddc80577f010d1a291ec26b" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.219199 5002 generic.go:334] "Generic (PLEG): container finished" podID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerID="542d9dc966b99803bafeed52ab3f95e09b26fa4621c3c7ba2b413b0a920b7ab9" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.219226 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerDied","Data":"1717340cfb5720d63c8157ee059c297fe8ebbf217ddc80577f010d1a291ec26b"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.219286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerDied","Data":"542d9dc966b99803bafeed52ab3f95e09b26fa4621c3c7ba2b413b0a920b7ab9"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222191 5002 generic.go:334] "Generic (PLEG): container finished" podID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerID="995104a78065f95a3691c5e595209bf467b2d52e6a7ae0e3ccabca98bd6178e9" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222212 5002 generic.go:334] "Generic (PLEG): container finished" podID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerID="f303ff5fb8c6acb8296f9ead1a3e6a64ea6507b6cfbb4b864a1d46b3de3bd8db" exitCode=137 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222352 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-central-agent" containerID="cri-o://59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550" gracePeriod=30 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222628 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerDied","Data":"995104a78065f95a3691c5e595209bf467b2d52e6a7ae0e3ccabca98bd6178e9"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222655 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerDied","Data":"f303ff5fb8c6acb8296f9ead1a3e6a64ea6507b6cfbb4b864a1d46b3de3bd8db"} Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222935 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="sg-core" containerID="cri-o://7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2" gracePeriod=30 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.222987 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-notification-agent" containerID="cri-o://2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f" gracePeriod=30 Oct 14 08:08:12 crc kubenswrapper[5002]: I1014 08:08:12.224669 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.232658 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerID="7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2" exitCode=2 Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.232931 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerID="59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550" exitCode=0 Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.232890 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerDied","Data":"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2"} Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.232969 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerDied","Data":"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550"} Oct 14 08:08:13 crc kubenswrapper[5002]: E1014 08:08:13.625276 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429" Oct 14 08:08:13 crc kubenswrapper[5002]: E1014 08:08:13.625446 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mtj4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-v746c_openstack(0307318a-577d-4b25-9972-274b46d62f36): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:08:13 crc kubenswrapper[5002]: E1014 08:08:13.626608 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-v746c" podUID="0307318a-577d-4b25-9972-274b46d62f36" Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.952048 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.978781 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:08:13 crc kubenswrapper[5002]: I1014 08:08:13.986521 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-557d55dc9d-sk8l2" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.083732 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data\") pod \"12720d49-a4e4-426e-bf38-169a0beeb8d2\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.083798 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key\") pod \"12720d49-a4e4-426e-bf38-169a0beeb8d2\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.083855 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle\") pod \"8bbe80a8-92fa-4f55-9937-414e1755a05e\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.083933 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts\") pod \"12720d49-a4e4-426e-bf38-169a0beeb8d2\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.083982 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs\") pod \"12720d49-a4e4-426e-bf38-169a0beeb8d2\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.084047 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6xqn\" (UniqueName: \"kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn\") pod \"8bbe80a8-92fa-4f55-9937-414e1755a05e\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.084104 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config\") pod \"8bbe80a8-92fa-4f55-9937-414e1755a05e\" (UID: \"8bbe80a8-92fa-4f55-9937-414e1755a05e\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.084165 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhjff\" (UniqueName: \"kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff\") pod \"12720d49-a4e4-426e-bf38-169a0beeb8d2\" (UID: \"12720d49-a4e4-426e-bf38-169a0beeb8d2\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.094706 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.095070 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon-log" containerID="cri-o://ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990" gracePeriod=30 Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.095483 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" containerID="cri-o://b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2" gracePeriod=30 Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.095757 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs" (OuterVolumeSpecName: "logs") pod "12720d49-a4e4-426e-bf38-169a0beeb8d2" (UID: "12720d49-a4e4-426e-bf38-169a0beeb8d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.111074 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn" (OuterVolumeSpecName: "kube-api-access-q6xqn") pod "8bbe80a8-92fa-4f55-9937-414e1755a05e" (UID: "8bbe80a8-92fa-4f55-9937-414e1755a05e"). InnerVolumeSpecName "kube-api-access-q6xqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.112010 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "12720d49-a4e4-426e-bf38-169a0beeb8d2" (UID: "12720d49-a4e4-426e-bf38-169a0beeb8d2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.112222 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.151107 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts" (OuterVolumeSpecName: "scripts") pod "12720d49-a4e4-426e-bf38-169a0beeb8d2" (UID: "12720d49-a4e4-426e-bf38-169a0beeb8d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.152609 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff" (OuterVolumeSpecName: "kube-api-access-hhjff") pod "12720d49-a4e4-426e-bf38-169a0beeb8d2" (UID: "12720d49-a4e4-426e-bf38-169a0beeb8d2"). InnerVolumeSpecName "kube-api-access-hhjff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.177440 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config" (OuterVolumeSpecName: "config") pod "8bbe80a8-92fa-4f55-9937-414e1755a05e" (UID: "8bbe80a8-92fa-4f55-9937-414e1755a05e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.192658 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data" (OuterVolumeSpecName: "config-data") pod "12720d49-a4e4-426e-bf38-169a0beeb8d2" (UID: "12720d49-a4e4-426e-bf38-169a0beeb8d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.196950 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.196978 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12720d49-a4e4-426e-bf38-169a0beeb8d2-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.196988 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6xqn\" (UniqueName: \"kubernetes.io/projected/8bbe80a8-92fa-4f55-9937-414e1755a05e-kube-api-access-q6xqn\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.196998 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.197006 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhjff\" (UniqueName: \"kubernetes.io/projected/12720d49-a4e4-426e-bf38-169a0beeb8d2-kube-api-access-hhjff\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.197015 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/12720d49-a4e4-426e-bf38-169a0beeb8d2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.197027 5002 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/12720d49-a4e4-426e-bf38-169a0beeb8d2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.210137 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.218316 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bbe80a8-92fa-4f55-9937-414e1755a05e" (UID: "8bbe80a8-92fa-4f55-9937-414e1755a05e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.253595 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.255437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5dc898f-nqhn4" event={"ID":"12720d49-a4e4-426e-bf38-169a0beeb8d2","Type":"ContainerDied","Data":"df5f6319a166b0c7cf9152000b3df0631123ead38811d24faffc9654f437c6e6"} Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.255474 5002 scope.go:117] "RemoveContainer" containerID="783cb420ade93fc94cb405135847f53d76475fa05bb6ec01bbe8cac8778c707e" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.256552 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5dc898f-nqhn4" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.265865 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77cd4c5db5-24j69" event={"ID":"2e99bd9e-3960-439c-a053-f0ecdde86ccb","Type":"ContainerDied","Data":"671ba1175ccc5bb8f6779cc658eb5f5dfb6bd7c74f2e941581393bd8e545f2fe"} Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.265937 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77cd4c5db5-24j69" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.268797 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-577688b66f-svbn9" event={"ID":"6aa9bf96-1917-467e-98c7-7cbca1945278","Type":"ContainerDied","Data":"49aa15a0ced23cd81fd7acc3791150335541d212a15ea2475be209ec18843b56"} Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.268942 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-577688b66f-svbn9" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.270924 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-7wzf7" event={"ID":"8bbe80a8-92fa-4f55-9937-414e1755a05e","Type":"ContainerDied","Data":"d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf"} Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.271015 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9665c469daa5e64d09653cf2449b6ec60eafa6a1e33e5820e2619e23f267cdf" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.271104 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-7wzf7" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.278218 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fbcgt" event={"ID":"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69","Type":"ContainerStarted","Data":"efe005d5dad3f56bf17f8f278af3befbf433698afb2da4607320882132b0cbaf"} Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.303339 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs\") pod \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.303608 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x8s5\" (UniqueName: \"kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5\") pod \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.303688 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key\") pod \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.303744 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts\") pod \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.303812 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data\") pod \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\" (UID: \"2e99bd9e-3960-439c-a053-f0ecdde86ccb\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.304196 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbe80a8-92fa-4f55-9937-414e1755a05e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.306537 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs" (OuterVolumeSpecName: "logs") pod "2e99bd9e-3960-439c-a053-f0ecdde86ccb" (UID: "2e99bd9e-3960-439c-a053-f0ecdde86ccb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.325756 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-fbcgt" podStartSLOduration=7.260669706 podStartE2EDuration="28.325738375s" podCreationTimestamp="2025-10-14 08:07:46 +0000 UTC" firstStartedPulling="2025-10-14 08:07:52.519965871 +0000 UTC m=+1005.501205323" lastFinishedPulling="2025-10-14 08:08:13.58503454 +0000 UTC m=+1026.566273992" observedRunningTime="2025-10-14 08:08:14.307346681 +0000 UTC m=+1027.288586133" watchObservedRunningTime="2025-10-14 08:08:14.325738375 +0000 UTC m=+1027.306977827" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.339189 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5" (OuterVolumeSpecName: "kube-api-access-4x8s5") pod "2e99bd9e-3960-439c-a053-f0ecdde86ccb" (UID: "2e99bd9e-3960-439c-a053-f0ecdde86ccb"). InnerVolumeSpecName "kube-api-access-4x8s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.344417 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2e99bd9e-3960-439c-a053-f0ecdde86ccb" (UID: "2e99bd9e-3960-439c-a053-f0ecdde86ccb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.348511 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data" (OuterVolumeSpecName: "config-data") pod "2e99bd9e-3960-439c-a053-f0ecdde86ccb" (UID: "2e99bd9e-3960-439c-a053-f0ecdde86ccb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.361228 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts" (OuterVolumeSpecName: "scripts") pod "2e99bd9e-3960-439c-a053-f0ecdde86ccb" (UID: "2e99bd9e-3960-439c-a053-f0ecdde86ccb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.368043 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.374740 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d5dc898f-nqhn4"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405165 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs\") pod \"6aa9bf96-1917-467e-98c7-7cbca1945278\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405203 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data\") pod \"6aa9bf96-1917-467e-98c7-7cbca1945278\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405386 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key\") pod \"6aa9bf96-1917-467e-98c7-7cbca1945278\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405419 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrlwb\" (UniqueName: \"kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb\") pod \"6aa9bf96-1917-467e-98c7-7cbca1945278\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405490 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts\") pod \"6aa9bf96-1917-467e-98c7-7cbca1945278\" (UID: \"6aa9bf96-1917-467e-98c7-7cbca1945278\") " Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.405819 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs" (OuterVolumeSpecName: "logs") pod "6aa9bf96-1917-467e-98c7-7cbca1945278" (UID: "6aa9bf96-1917-467e-98c7-7cbca1945278"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406446 5002 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e99bd9e-3960-439c-a053-f0ecdde86ccb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406683 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406699 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e99bd9e-3960-439c-a053-f0ecdde86ccb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406718 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa9bf96-1917-467e-98c7-7cbca1945278-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406727 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e99bd9e-3960-439c-a053-f0ecdde86ccb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.406736 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x8s5\" (UniqueName: \"kubernetes.io/projected/2e99bd9e-3960-439c-a053-f0ecdde86ccb-kube-api-access-4x8s5\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.408710 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb" (OuterVolumeSpecName: "kube-api-access-lrlwb") pod "6aa9bf96-1917-467e-98c7-7cbca1945278" (UID: "6aa9bf96-1917-467e-98c7-7cbca1945278"). InnerVolumeSpecName "kube-api-access-lrlwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.409697 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6aa9bf96-1917-467e-98c7-7cbca1945278" (UID: "6aa9bf96-1917-467e-98c7-7cbca1945278"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.426611 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts" (OuterVolumeSpecName: "scripts") pod "6aa9bf96-1917-467e-98c7-7cbca1945278" (UID: "6aa9bf96-1917-467e-98c7-7cbca1945278"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.427630 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data" (OuterVolumeSpecName: "config-data") pod "6aa9bf96-1917-467e-98c7-7cbca1945278" (UID: "6aa9bf96-1917-467e-98c7-7cbca1945278"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.437484 5002 scope.go:117] "RemoveContainer" containerID="6713e666d17a7b8c0557c873580a43ff3fd96b008965abade3601b40a1398a83" Oct 14 08:08:14 crc kubenswrapper[5002]: E1014 08:08:14.437679 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429\\\"\"" pod="openstack/cinder-db-sync-v746c" podUID="0307318a-577d-4b25-9972-274b46d62f36" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.468992 5002 scope.go:117] "RemoveContainer" containerID="995104a78065f95a3691c5e595209bf467b2d52e6a7ae0e3ccabca98bd6178e9" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.507942 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.507978 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6aa9bf96-1917-467e-98c7-7cbca1945278-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.507989 5002 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6aa9bf96-1917-467e-98c7-7cbca1945278-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.508000 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrlwb\" (UniqueName: \"kubernetes.io/projected/6aa9bf96-1917-467e-98c7-7cbca1945278-kube-api-access-lrlwb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.636169 5002 scope.go:117] "RemoveContainer" containerID="f303ff5fb8c6acb8296f9ead1a3e6a64ea6507b6cfbb4b864a1d46b3de3bd8db" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.700474 5002 scope.go:117] "RemoveContainer" containerID="1717340cfb5720d63c8157ee059c297fe8ebbf217ddc80577f010d1a291ec26b" Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.702669 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.710378 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-577688b66f-svbn9"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.717363 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.724228 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77cd4c5db5-24j69"] Oct 14 08:08:14 crc kubenswrapper[5002]: I1014 08:08:14.847050 5002 scope.go:117] "RemoveContainer" containerID="542d9dc966b99803bafeed52ab3f95e09b26fa4621c3c7ba2b413b0a920b7ab9" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.233996 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234690 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234705 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234717 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234723 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234733 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234742 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234759 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234766 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234781 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234787 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234801 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234807 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: E1014 08:08:15.234856 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbe80a8-92fa-4f55-9937-414e1755a05e" containerName="neutron-db-sync" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.234864 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbe80a8-92fa-4f55-9937-414e1755a05e" containerName="neutron-db-sync" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235043 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235063 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235083 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon-log" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235094 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235110 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235121 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbe80a8-92fa-4f55-9937-414e1755a05e" containerName="neutron-db-sync" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.235137 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" containerName="horizon" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.241205 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.244585 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.297944 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.299199 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.309912 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.309980 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.310090 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vlswn" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.310267 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.317216 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.326226 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.326253 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk2tw\" (UniqueName: \"kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.326286 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.326384 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.326413 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.428801 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.428860 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk2tw\" (UniqueName: \"kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.428907 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.428941 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.428978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429059 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429077 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mglr\" (UniqueName: \"kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429098 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429128 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429155 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429672 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.429863 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.430550 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.431601 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.448729 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk2tw\" (UniqueName: \"kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw\") pod \"dnsmasq-dns-56ff5947d9-tcrzr\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.530497 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.530539 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mglr\" (UniqueName: \"kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.530563 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.530654 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.530683 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.548112 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.554522 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.554684 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.555287 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.555460 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mglr\" (UniqueName: \"kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr\") pod \"neutron-7cfc656d-2j56s\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.566001 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.634809 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.736035 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12720d49-a4e4-426e-bf38-169a0beeb8d2" path="/var/lib/kubelet/pods/12720d49-a4e4-426e-bf38-169a0beeb8d2/volumes" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.737130 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e99bd9e-3960-439c-a053-f0ecdde86ccb" path="/var/lib/kubelet/pods/2e99bd9e-3960-439c-a053-f0ecdde86ccb/volumes" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.738768 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa9bf96-1917-467e-98c7-7cbca1945278" path="/var/lib/kubelet/pods/6aa9bf96-1917-467e-98c7-7cbca1945278/volumes" Oct 14 08:08:15 crc kubenswrapper[5002]: I1014 08:08:15.910539 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.040901 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.040988 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041011 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041036 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041159 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppgq8\" (UniqueName: \"kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041190 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041241 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd\") pod \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\" (UID: \"7a5a0fa6-e889-4d56-b2c7-070aa5012a10\") " Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041320 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041535 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.041754 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.045641 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts" (OuterVolumeSpecName: "scripts") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.045913 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8" (OuterVolumeSpecName: "kube-api-access-ppgq8") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "kube-api-access-ppgq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.098029 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.125994 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.144196 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.144232 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.144245 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.144256 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppgq8\" (UniqueName: \"kubernetes.io/projected/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-kube-api-access-ppgq8\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.144267 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.145057 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data" (OuterVolumeSpecName: "config-data") pod "7a5a0fa6-e889-4d56-b2c7-070aa5012a10" (UID: "7a5a0fa6-e889-4d56-b2c7-070aa5012a10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.173204 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.246099 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a5a0fa6-e889-4d56-b2c7-070aa5012a10-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.341206 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerID="2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f" exitCode=0 Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.341256 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerDied","Data":"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f"} Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.341324 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7a5a0fa6-e889-4d56-b2c7-070aa5012a10","Type":"ContainerDied","Data":"317afa8bd12836ac2bcc3588f6613edd32c232146b1f149de98c5938c7acde9c"} Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.341320 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.341345 5002 scope.go:117] "RemoveContainer" containerID="7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.350966 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerStarted","Data":"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe"} Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.351021 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerStarted","Data":"6042edb545c99418e2c85fdbdbb4da1a95adb88cbc4cadb5d7c5fff13e6b6792"} Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.362156 5002 scope.go:117] "RemoveContainer" containerID="2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.419155 5002 scope.go:117] "RemoveContainer" containerID="59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.447513 5002 scope.go:117] "RemoveContainer" containerID="7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2" Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.450449 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2\": container with ID starting with 7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2 not found: ID does not exist" containerID="7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.450498 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2"} err="failed to get container status \"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2\": rpc error: code = NotFound desc = could not find container \"7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2\": container with ID starting with 7b2d882dd3f3e779e7d6b40c7348fafd5e574cd2e0706f368da6b5ac9458e5b2 not found: ID does not exist" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.450530 5002 scope.go:117] "RemoveContainer" containerID="2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.450638 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.451259 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f\": container with ID starting with 2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f not found: ID does not exist" containerID="2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.451299 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f"} err="failed to get container status \"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f\": rpc error: code = NotFound desc = could not find container \"2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f\": container with ID starting with 2e0e49ddfe38614c4fe1a8e0f54913f3c6d7b1ae525871465e8efd793903f84f not found: ID does not exist" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.451326 5002 scope.go:117] "RemoveContainer" containerID="59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550" Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.451661 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550\": container with ID starting with 59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550 not found: ID does not exist" containerID="59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.451683 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550"} err="failed to get container status \"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550\": rpc error: code = NotFound desc = could not find container \"59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550\": container with ID starting with 59d16c27b40c70f909e5c2faaa50a0dececef3916412632b849c882c304b4550 not found: ID does not exist" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.460813 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.477213 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.477695 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-notification-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.477721 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-notification-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.477732 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="sg-core" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.477740 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="sg-core" Oct 14 08:08:16 crc kubenswrapper[5002]: E1014 08:08:16.477749 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-central-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.477756 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-central-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.486429 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="sg-core" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.486533 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-notification-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.486556 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" containerName="ceilometer-central-agent" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.488798 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.488906 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.491728 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.492389 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553529 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553665 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553717 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553862 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553911 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.553939 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5gtv\" (UniqueName: \"kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.554048 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.641739 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:16 crc kubenswrapper[5002]: W1014 08:08:16.648281 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7091e2e_4bcf_48e7_814a_39179fad09d4.slice/crio-119f1afe824d262b3b54e7c5c1608242b77e9ee11457880bed8ac6439eaf1ba0 WatchSource:0}: Error finding container 119f1afe824d262b3b54e7c5c1608242b77e9ee11457880bed8ac6439eaf1ba0: Status 404 returned error can't find the container with id 119f1afe824d262b3b54e7c5c1608242b77e9ee11457880bed8ac6439eaf1ba0 Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.663846 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.663946 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.663983 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664003 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664030 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664055 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664072 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5gtv\" (UniqueName: \"kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664351 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.664451 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.667794 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.668101 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.668541 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.675532 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.687793 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5gtv\" (UniqueName: \"kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv\") pod \"ceilometer-0\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " pod="openstack/ceilometer-0" Oct 14 08:08:16 crc kubenswrapper[5002]: I1014 08:08:16.805225 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.241869 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:33172->10.217.0.140:8443: read: connection reset by peer" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.324915 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:17 crc kubenswrapper[5002]: W1014 08:08:17.343661 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod866e36d1_d1ee_49c8_8052_fa5097ca695b.slice/crio-d8c8bc26c7664de28159988f5e4f6b038f9f4671ab88a72f0099407690a419bd WatchSource:0}: Error finding container d8c8bc26c7664de28159988f5e4f6b038f9f4671ab88a72f0099407690a419bd: Status 404 returned error can't find the container with id d8c8bc26c7664de28159988f5e4f6b038f9f4671ab88a72f0099407690a419bd Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.359207 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerStarted","Data":"d8c8bc26c7664de28159988f5e4f6b038f9f4671ab88a72f0099407690a419bd"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.361428 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerStarted","Data":"69ae80016f8c5320fc50b1b9bce8c35caa9dcd2528dcd0d705d082389b37e845"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.361463 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerStarted","Data":"9e71e0cf8fbe795bdbcf8d98fcb033b339f5649ea1fe370d253c12cd71a8129b"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.361479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerStarted","Data":"119f1afe824d262b3b54e7c5c1608242b77e9ee11457880bed8ac6439eaf1ba0"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.361586 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.364461 5002 generic.go:334] "Generic (PLEG): container finished" podID="d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" containerID="efe005d5dad3f56bf17f8f278af3befbf433698afb2da4607320882132b0cbaf" exitCode=0 Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.364548 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fbcgt" event={"ID":"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69","Type":"ContainerDied","Data":"efe005d5dad3f56bf17f8f278af3befbf433698afb2da4607320882132b0cbaf"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.370362 5002 generic.go:334] "Generic (PLEG): container finished" podID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerID="b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2" exitCode=0 Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.370413 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerDied","Data":"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.372698 5002 generic.go:334] "Generic (PLEG): container finished" podID="056e91f0-af24-487c-998e-55519661448d" containerID="7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe" exitCode=0 Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.372747 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerDied","Data":"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe"} Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.386739 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7cfc656d-2j56s" podStartSLOduration=2.386718977 podStartE2EDuration="2.386718977s" podCreationTimestamp="2025-10-14 08:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:17.380369366 +0000 UTC m=+1030.361608828" watchObservedRunningTime="2025-10-14 08:08:17.386718977 +0000 UTC m=+1030.367958439" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.729915 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a5a0fa6-e889-4d56-b2c7-070aa5012a10" path="/var/lib/kubelet/pods/7a5a0fa6-e889-4d56-b2c7-070aa5012a10/volumes" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.799000 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7855f7b865-dtwlh"] Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.800741 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.804977 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.809996 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.833948 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7855f7b865-dtwlh"] Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.885386 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-public-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.885821 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bn4h\" (UniqueName: \"kubernetes.io/projected/abe829e6-d42f-447f-bcb2-aad97c1e4f92-kube-api-access-9bn4h\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.885987 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-combined-ca-bundle\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.886123 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.886312 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-ovndb-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.886405 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-internal-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.886802 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-httpd-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.987943 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-public-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988316 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bn4h\" (UniqueName: \"kubernetes.io/projected/abe829e6-d42f-447f-bcb2-aad97c1e4f92-kube-api-access-9bn4h\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988359 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-combined-ca-bundle\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988419 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988491 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-ovndb-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988520 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-internal-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.988556 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-httpd-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.992695 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-httpd-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.992823 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-public-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.992917 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-config\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:17 crc kubenswrapper[5002]: I1014 08:08:17.994017 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-internal-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:17.997183 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-ovndb-tls-certs\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.006604 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe829e6-d42f-447f-bcb2-aad97c1e4f92-combined-ca-bundle\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.013652 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bn4h\" (UniqueName: \"kubernetes.io/projected/abe829e6-d42f-447f-bcb2-aad97c1e4f92-kube-api-access-9bn4h\") pod \"neutron-7855f7b865-dtwlh\" (UID: \"abe829e6-d42f-447f-bcb2-aad97c1e4f92\") " pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.122796 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.393340 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerStarted","Data":"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e"} Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.417250 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerStarted","Data":"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351"} Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.417790 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.436215 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" podStartSLOduration=3.436200162 podStartE2EDuration="3.436200162s" podCreationTimestamp="2025-10-14 08:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:18.434531627 +0000 UTC m=+1031.415771089" watchObservedRunningTime="2025-10-14 08:08:18.436200162 +0000 UTC m=+1031.417439614" Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.475478 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7855f7b865-dtwlh"] Oct 14 08:08:18 crc kubenswrapper[5002]: I1014 08:08:18.878786 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.005485 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdqxb\" (UniqueName: \"kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb\") pod \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.005534 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data\") pod \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.005569 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle\") pod \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\" (UID: \"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69\") " Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.013480 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" (UID: "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.018134 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb" (OuterVolumeSpecName: "kube-api-access-sdqxb") pod "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" (UID: "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69"). InnerVolumeSpecName "kube-api-access-sdqxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.057056 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" (UID: "d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.107838 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdqxb\" (UniqueName: \"kubernetes.io/projected/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-kube-api-access-sdqxb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.108119 5002 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.108128 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.447241 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fbcgt" event={"ID":"d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69","Type":"ContainerDied","Data":"31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.447280 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c7e56edc66085c340a8e5680fbb5e83c857b797c60ae35696f361d32f0e4da" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.447568 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fbcgt" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.448709 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7855f7b865-dtwlh" event={"ID":"abe829e6-d42f-447f-bcb2-aad97c1e4f92","Type":"ContainerStarted","Data":"5be2417b91f685c0630b4c057a6fbcdfed1c7e2369810dad48a3318791d77ae8"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.448736 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7855f7b865-dtwlh" event={"ID":"abe829e6-d42f-447f-bcb2-aad97c1e4f92","Type":"ContainerStarted","Data":"b3c71e12ed5a06c63276860fa200cc2098a9af52f746fd102790118ace92424b"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.448750 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7855f7b865-dtwlh" event={"ID":"abe829e6-d42f-447f-bcb2-aad97c1e4f92","Type":"ContainerStarted","Data":"abb00a177867f0ce38570bc42f3124a9d4e86b89575e04a7b114bd76c5d9160e"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.449828 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.462575 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerStarted","Data":"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.462634 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerStarted","Data":"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe"} Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.484656 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7855f7b865-dtwlh" podStartSLOduration=2.48463031 podStartE2EDuration="2.48463031s" podCreationTimestamp="2025-10-14 08:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:19.481406634 +0000 UTC m=+1032.462646096" watchObservedRunningTime="2025-10-14 08:08:19.48463031 +0000 UTC m=+1032.465869762" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.578658 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7fffd76679-n928f"] Oct 14 08:08:19 crc kubenswrapper[5002]: E1014 08:08:19.578990 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" containerName="barbican-db-sync" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.579008 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" containerName="barbican-db-sync" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.579182 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" containerName="barbican-db-sync" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.580017 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.582177 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rtxxg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.582447 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.591169 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.592495 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7fffd76679-n928f"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.619392 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7f846844d6-pvmrs"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.620745 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.625074 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.637806 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f846844d6-pvmrs"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.717524 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721725 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data-custom\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721779 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-combined-ca-bundle\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721820 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-combined-ca-bundle\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721848 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data-custom\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721915 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721939 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j65mz\" (UniqueName: \"kubernetes.io/projected/a1f84982-1a7e-4054-b3f5-666de26a4e96-kube-api-access-j65mz\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721955 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec0b4c68-f07d-4d42-8225-6950e9315970-logs\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.721986 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpz8p\" (UniqueName: \"kubernetes.io/projected/ec0b4c68-f07d-4d42-8225-6950e9315970-kube-api-access-hpz8p\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.722014 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f84982-1a7e-4054-b3f5-666de26a4e96-logs\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.722032 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.745165 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.748344 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.762239 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.792236 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.794677 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.797432 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.814218 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.822987 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data-custom\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.827994 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-combined-ca-bundle\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828044 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data-custom\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828133 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828179 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-combined-ca-bundle\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828221 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data-custom\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828341 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828381 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828420 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j65mz\" (UniqueName: \"kubernetes.io/projected/a1f84982-1a7e-4054-b3f5-666de26a4e96-kube-api-access-j65mz\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828444 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec0b4c68-f07d-4d42-8225-6950e9315970-logs\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828489 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828520 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpz8p\" (UniqueName: \"kubernetes.io/projected/ec0b4c68-f07d-4d42-8225-6950e9315970-kube-api-access-hpz8p\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828569 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828596 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f84982-1a7e-4054-b3f5-666de26a4e96-logs\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828617 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.828644 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwl7v\" (UniqueName: \"kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.830502 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec0b4c68-f07d-4d42-8225-6950e9315970-logs\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.832757 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-combined-ca-bundle\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.837721 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-combined-ca-bundle\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.838357 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data-custom\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.843106 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f84982-1a7e-4054-b3f5-666de26a4e96-config-data\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.844042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec0b4c68-f07d-4d42-8225-6950e9315970-config-data\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.844095 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f84982-1a7e-4054-b3f5-666de26a4e96-logs\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.861166 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpz8p\" (UniqueName: \"kubernetes.io/projected/ec0b4c68-f07d-4d42-8225-6950e9315970-kube-api-access-hpz8p\") pod \"barbican-worker-7fffd76679-n928f\" (UID: \"ec0b4c68-f07d-4d42-8225-6950e9315970\") " pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.865022 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j65mz\" (UniqueName: \"kubernetes.io/projected/a1f84982-1a7e-4054-b3f5-666de26a4e96-kube-api-access-j65mz\") pod \"barbican-keystone-listener-7f846844d6-pvmrs\" (UID: \"a1f84982-1a7e-4054-b3f5-666de26a4e96\") " pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.901983 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.904825 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7fffd76679-n928f" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930286 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930344 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930369 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930397 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qwlp\" (UniqueName: \"kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930449 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930486 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930510 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930526 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930560 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.930583 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwl7v\" (UniqueName: \"kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.931633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.932161 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.933045 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.933410 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.946575 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwl7v\" (UniqueName: \"kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v\") pod \"dnsmasq-dns-85f7955b85-fm7pg\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:19 crc kubenswrapper[5002]: I1014 08:08:19.970151 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.035245 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.035291 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.036246 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.036548 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.036638 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qwlp\" (UniqueName: \"kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.037493 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.043491 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.043633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.043997 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.056683 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qwlp\" (UniqueName: \"kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp\") pod \"barbican-api-74b55bb47b-hbxr2\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.065312 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.125538 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.350279 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7fffd76679-n928f"] Oct 14 08:08:20 crc kubenswrapper[5002]: W1014 08:08:20.363813 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec0b4c68_f07d_4d42_8225_6950e9315970.slice/crio-8a19aa7e29d36eb66579b20679c2b4803e850671f871930b81693368084e8035 WatchSource:0}: Error finding container 8a19aa7e29d36eb66579b20679c2b4803e850671f871930b81693368084e8035: Status 404 returned error can't find the container with id 8a19aa7e29d36eb66579b20679c2b4803e850671f871930b81693368084e8035 Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.464591 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7fffd76679-n928f" event={"ID":"ec0b4c68-f07d-4d42-8225-6950e9315970","Type":"ContainerStarted","Data":"8a19aa7e29d36eb66579b20679c2b4803e850671f871930b81693368084e8035"} Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.464719 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="dnsmasq-dns" containerID="cri-o://021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351" gracePeriod=10 Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.499315 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7f846844d6-pvmrs"] Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.613640 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:20 crc kubenswrapper[5002]: W1014 08:08:20.636911 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb33216f_b3e7_418f_ac05_129fb34f4538.slice/crio-c66098917b50f823128ff878ae39db3f8bbb1b3002b8d155f6b8591b1252170b WatchSource:0}: Error finding container c66098917b50f823128ff878ae39db3f8bbb1b3002b8d155f6b8591b1252170b: Status 404 returned error can't find the container with id c66098917b50f823128ff878ae39db3f8bbb1b3002b8d155f6b8591b1252170b Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.696506 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.878883 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.959813 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc\") pod \"056e91f0-af24-487c-998e-55519661448d\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.960150 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb\") pod \"056e91f0-af24-487c-998e-55519661448d\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.960323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb\") pod \"056e91f0-af24-487c-998e-55519661448d\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.960367 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config\") pod \"056e91f0-af24-487c-998e-55519661448d\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.960391 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk2tw\" (UniqueName: \"kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw\") pod \"056e91f0-af24-487c-998e-55519661448d\" (UID: \"056e91f0-af24-487c-998e-55519661448d\") " Oct 14 08:08:20 crc kubenswrapper[5002]: I1014 08:08:20.966914 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw" (OuterVolumeSpecName: "kube-api-access-tk2tw") pod "056e91f0-af24-487c-998e-55519661448d" (UID: "056e91f0-af24-487c-998e-55519661448d"). InnerVolumeSpecName "kube-api-access-tk2tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.036381 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "056e91f0-af24-487c-998e-55519661448d" (UID: "056e91f0-af24-487c-998e-55519661448d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.064436 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk2tw\" (UniqueName: \"kubernetes.io/projected/056e91f0-af24-487c-998e-55519661448d-kube-api-access-tk2tw\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.064461 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.068798 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "056e91f0-af24-487c-998e-55519661448d" (UID: "056e91f0-af24-487c-998e-55519661448d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.075347 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "056e91f0-af24-487c-998e-55519661448d" (UID: "056e91f0-af24-487c-998e-55519661448d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.078008 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config" (OuterVolumeSpecName: "config") pod "056e91f0-af24-487c-998e-55519661448d" (UID: "056e91f0-af24-487c-998e-55519661448d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.166333 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.166368 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.166378 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056e91f0-af24-487c-998e-55519661448d-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.474947 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" event={"ID":"a1f84982-1a7e-4054-b3f5-666de26a4e96","Type":"ContainerStarted","Data":"2ca15269a36a729533a57491a060fc1f5a1358e44b7c323e2c597e515620c7fe"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.478089 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerStarted","Data":"5cab479b92c0dc5af912937d80d978e5dd7c1f03d9435b3e06270666b537f991"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.478132 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerStarted","Data":"602a385670a710c0f2a0160b6dc7a2325c485a720d8642bba52663af13f990aa"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.478141 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerStarted","Data":"82e1813334b0e56d814271a9fee5947a00216ee43737c89871617d81e63c33c3"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.479179 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.479207 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.484435 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerStarted","Data":"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.484551 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.490161 5002 generic.go:334] "Generic (PLEG): container finished" podID="db33216f-b3e7-418f-ac05-129fb34f4538" containerID="b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624" exitCode=0 Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.490225 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" event={"ID":"db33216f-b3e7-418f-ac05-129fb34f4538","Type":"ContainerDied","Data":"b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.490297 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" event={"ID":"db33216f-b3e7-418f-ac05-129fb34f4538","Type":"ContainerStarted","Data":"c66098917b50f823128ff878ae39db3f8bbb1b3002b8d155f6b8591b1252170b"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.498416 5002 generic.go:334] "Generic (PLEG): container finished" podID="056e91f0-af24-487c-998e-55519661448d" containerID="021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351" exitCode=0 Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.498465 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.498521 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerDied","Data":"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.498549 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56ff5947d9-tcrzr" event={"ID":"056e91f0-af24-487c-998e-55519661448d","Type":"ContainerDied","Data":"6042edb545c99418e2c85fdbdbb4da1a95adb88cbc4cadb5d7c5fff13e6b6792"} Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.498564 5002 scope.go:117] "RemoveContainer" containerID="021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.531645 5002 scope.go:117] "RemoveContainer" containerID="7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.582945 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74b55bb47b-hbxr2" podStartSLOduration=2.582925854 podStartE2EDuration="2.582925854s" podCreationTimestamp="2025-10-14 08:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:21.519585325 +0000 UTC m=+1034.500824777" watchObservedRunningTime="2025-10-14 08:08:21.582925854 +0000 UTC m=+1034.564165306" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.586558 5002 scope.go:117] "RemoveContainer" containerID="021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351" Oct 14 08:08:21 crc kubenswrapper[5002]: E1014 08:08:21.589241 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351\": container with ID starting with 021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351 not found: ID does not exist" containerID="021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.589318 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351"} err="failed to get container status \"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351\": rpc error: code = NotFound desc = could not find container \"021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351\": container with ID starting with 021009e8a67c99062366692b33d9be4affbc1ebe402f971849c07fa5b75d5351 not found: ID does not exist" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.589344 5002 scope.go:117] "RemoveContainer" containerID="7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe" Oct 14 08:08:21 crc kubenswrapper[5002]: E1014 08:08:21.593969 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe\": container with ID starting with 7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe not found: ID does not exist" containerID="7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.594011 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe"} err="failed to get container status \"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe\": rpc error: code = NotFound desc = could not find container \"7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe\": container with ID starting with 7869bea473d75a233b214a4478395cdac1d33684f3aa60f14512b2921c6e2abe not found: ID does not exist" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.601748 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.416071682 podStartE2EDuration="5.601728308s" podCreationTimestamp="2025-10-14 08:08:16 +0000 UTC" firstStartedPulling="2025-10-14 08:08:17.346933259 +0000 UTC m=+1030.328172701" lastFinishedPulling="2025-10-14 08:08:20.532589875 +0000 UTC m=+1033.513829327" observedRunningTime="2025-10-14 08:08:21.57236362 +0000 UTC m=+1034.553603092" watchObservedRunningTime="2025-10-14 08:08:21.601728308 +0000 UTC m=+1034.582967760" Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.759588 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:21 crc kubenswrapper[5002]: I1014 08:08:21.759851 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56ff5947d9-tcrzr"] Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.146856 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-67d6789ffb-z5pvg"] Oct 14 08:08:22 crc kubenswrapper[5002]: E1014 08:08:22.151402 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="init" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.151423 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="init" Oct 14 08:08:22 crc kubenswrapper[5002]: E1014 08:08:22.151438 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="dnsmasq-dns" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.151444 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="dnsmasq-dns" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.151635 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="056e91f0-af24-487c-998e-55519661448d" containerName="dnsmasq-dns" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.152810 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.155993 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.164682 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.165315 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67d6789ffb-z5pvg"] Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.236756 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-public-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237001 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-logs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237079 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data-custom\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237178 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kf5f\" (UniqueName: \"kubernetes.io/projected/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-kube-api-access-2kf5f\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237255 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-combined-ca-bundle\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237336 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.237438 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-internal-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.339937 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-combined-ca-bundle\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.339999 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.340062 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-internal-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.340118 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-public-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.340149 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-logs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.340172 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data-custom\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.340220 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kf5f\" (UniqueName: \"kubernetes.io/projected/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-kube-api-access-2kf5f\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.341071 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-logs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.345450 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-public-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.345927 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.348352 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-internal-tls-certs\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.352755 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-combined-ca-bundle\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.353714 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-config-data-custom\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.356892 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kf5f\" (UniqueName: \"kubernetes.io/projected/e3b04e6e-b1f7-4327-985d-6ed4e526eeaa-kube-api-access-2kf5f\") pod \"barbican-api-67d6789ffb-z5pvg\" (UID: \"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa\") " pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.471541 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.525162 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" event={"ID":"db33216f-b3e7-418f-ac05-129fb34f4538","Type":"ContainerStarted","Data":"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7"} Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.525337 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:22 crc kubenswrapper[5002]: I1014 08:08:22.547955 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" podStartSLOduration=3.547935164 podStartE2EDuration="3.547935164s" podCreationTimestamp="2025-10-14 08:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:22.545982412 +0000 UTC m=+1035.527221884" watchObservedRunningTime="2025-10-14 08:08:22.547935164 +0000 UTC m=+1035.529174616" Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.350189 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-67d6789ffb-z5pvg"] Oct 14 08:08:23 crc kubenswrapper[5002]: W1014 08:08:23.361455 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3b04e6e_b1f7_4327_985d_6ed4e526eeaa.slice/crio-7ba6ab5f667c1f50d66710d7ed09bab2e158844503ed540c3ed1d091c5e14c6d WatchSource:0}: Error finding container 7ba6ab5f667c1f50d66710d7ed09bab2e158844503ed540c3ed1d091c5e14c6d: Status 404 returned error can't find the container with id 7ba6ab5f667c1f50d66710d7ed09bab2e158844503ed540c3ed1d091c5e14c6d Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.537679 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67d6789ffb-z5pvg" event={"ID":"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa","Type":"ContainerStarted","Data":"7ba6ab5f667c1f50d66710d7ed09bab2e158844503ed540c3ed1d091c5e14c6d"} Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.549983 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" event={"ID":"a1f84982-1a7e-4054-b3f5-666de26a4e96","Type":"ContainerStarted","Data":"d77dcdf35410c97587319f2c15ed7af502622955ad7c5702e5b7b65fe6ee8b19"} Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.550025 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" event={"ID":"a1f84982-1a7e-4054-b3f5-666de26a4e96","Type":"ContainerStarted","Data":"7341b1008bb5acf756d8f69697278d62257abc74292a1e6bcdd47cb99a33a422"} Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.553229 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7fffd76679-n928f" event={"ID":"ec0b4c68-f07d-4d42-8225-6950e9315970","Type":"ContainerStarted","Data":"957069b0485b60ddf5ea51bf8626717224929e0f22698afa6ea2d7262c936b36"} Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.553259 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7fffd76679-n928f" event={"ID":"ec0b4c68-f07d-4d42-8225-6950e9315970","Type":"ContainerStarted","Data":"c4d3e880a1a3d3e4ae6f946a9b3408d5a31f0a7dde8a974fefe437075ffcab0f"} Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.573461 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7f846844d6-pvmrs" podStartSLOduration=2.279410673 podStartE2EDuration="4.573443726s" podCreationTimestamp="2025-10-14 08:08:19 +0000 UTC" firstStartedPulling="2025-10-14 08:08:20.576694108 +0000 UTC m=+1033.557933570" lastFinishedPulling="2025-10-14 08:08:22.870727171 +0000 UTC m=+1035.851966623" observedRunningTime="2025-10-14 08:08:23.570134438 +0000 UTC m=+1036.551373900" watchObservedRunningTime="2025-10-14 08:08:23.573443726 +0000 UTC m=+1036.554683178" Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.592216 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7fffd76679-n928f" podStartSLOduration=2.087659721 podStartE2EDuration="4.59219688s" podCreationTimestamp="2025-10-14 08:08:19 +0000 UTC" firstStartedPulling="2025-10-14 08:08:20.365597276 +0000 UTC m=+1033.346836728" lastFinishedPulling="2025-10-14 08:08:22.870134435 +0000 UTC m=+1035.851373887" observedRunningTime="2025-10-14 08:08:23.588137291 +0000 UTC m=+1036.569376733" watchObservedRunningTime="2025-10-14 08:08:23.59219688 +0000 UTC m=+1036.573436332" Oct 14 08:08:23 crc kubenswrapper[5002]: I1014 08:08:23.729475 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="056e91f0-af24-487c-998e-55519661448d" path="/var/lib/kubelet/pods/056e91f0-af24-487c-998e-55519661448d/volumes" Oct 14 08:08:24 crc kubenswrapper[5002]: I1014 08:08:24.561394 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67d6789ffb-z5pvg" event={"ID":"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa","Type":"ContainerStarted","Data":"56fe4bf32c8cb27c9d995a6d11591df3f249093f22a6f369900df0178434ee4d"} Oct 14 08:08:24 crc kubenswrapper[5002]: I1014 08:08:24.561767 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-67d6789ffb-z5pvg" event={"ID":"e3b04e6e-b1f7-4327-985d-6ed4e526eeaa","Type":"ContainerStarted","Data":"404be494a79870b94f510add58aa22473e4a617cc7f54eb1be06c6a79c777710"} Oct 14 08:08:24 crc kubenswrapper[5002]: I1014 08:08:24.589122 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-67d6789ffb-z5pvg" podStartSLOduration=2.589101426 podStartE2EDuration="2.589101426s" podCreationTimestamp="2025-10-14 08:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:24.57847098 +0000 UTC m=+1037.559710462" watchObservedRunningTime="2025-10-14 08:08:24.589101426 +0000 UTC m=+1037.570340878" Oct 14 08:08:25 crc kubenswrapper[5002]: I1014 08:08:25.064692 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7db98f685f-fphzv" Oct 14 08:08:25 crc kubenswrapper[5002]: I1014 08:08:25.567907 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:25 crc kubenswrapper[5002]: I1014 08:08:25.568233 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:25 crc kubenswrapper[5002]: E1014 08:08:25.722285 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c\\\"\"" pod="openstack/placement-db-sync-jv98l" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" Oct 14 08:08:28 crc kubenswrapper[5002]: I1014 08:08:28.591018 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-v746c" event={"ID":"0307318a-577d-4b25-9972-274b46d62f36","Type":"ContainerStarted","Data":"c4b73f9c2443ba98ce446154d4c87a5d39eb6f2bfb951754a550de379f4c2a09"} Oct 14 08:08:28 crc kubenswrapper[5002]: I1014 08:08:28.612786 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-v746c" podStartSLOduration=7.895554662 podStartE2EDuration="42.612763875s" podCreationTimestamp="2025-10-14 08:07:46 +0000 UTC" firstStartedPulling="2025-10-14 08:07:52.529795284 +0000 UTC m=+1005.511034736" lastFinishedPulling="2025-10-14 08:08:27.247004497 +0000 UTC m=+1040.228243949" observedRunningTime="2025-10-14 08:08:28.607787602 +0000 UTC m=+1041.589027094" watchObservedRunningTime="2025-10-14 08:08:28.612763875 +0000 UTC m=+1041.594003337" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.300718 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.302145 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.308493 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.308778 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-mc99q" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.311518 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.312826 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.369415 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.369460 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.369513 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wljqq\" (UniqueName: \"kubernetes.io/projected/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-kube-api-access-wljqq\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.369805 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.471951 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.472026 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.472102 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wljqq\" (UniqueName: \"kubernetes.io/projected/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-kube-api-access-wljqq\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.472261 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.473323 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.479286 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-openstack-config-secret\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.481411 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.492055 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wljqq\" (UniqueName: \"kubernetes.io/projected/c0df2486-02ae-48b5-b3b9-acb14ec03a6f-kube-api-access-wljqq\") pod \"openstackclient\" (UID: \"c0df2486-02ae-48b5-b3b9-acb14ec03a6f\") " pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.625276 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:08:29 crc kubenswrapper[5002]: I1014 08:08:29.902566 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.068334 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.131959 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.132222 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="dnsmasq-dns" containerID="cri-o://46badaaa8912c6af3248be56592226d71d3ebc94ebd469fac2d66d7e695c6261" gracePeriod=10 Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.142487 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.624688 5002 generic.go:334] "Generic (PLEG): container finished" podID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerID="46badaaa8912c6af3248be56592226d71d3ebc94ebd469fac2d66d7e695c6261" exitCode=0 Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.624740 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" event={"ID":"e8a10d6e-f132-43fe-ac06-cf1a99c9146f","Type":"ContainerDied","Data":"46badaaa8912c6af3248be56592226d71d3ebc94ebd469fac2d66d7e695c6261"} Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.625103 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" event={"ID":"e8a10d6e-f132-43fe-ac06-cf1a99c9146f","Type":"ContainerDied","Data":"677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50"} Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.625123 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="677918c018ae806a7b64ad222ec12bf632ba8057c39c49ce4223bfb05bc65b50" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.626737 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c0df2486-02ae-48b5-b3b9-acb14ec03a6f","Type":"ContainerStarted","Data":"32d7d1f6b11bd795996bd0d570eaba7667fe4d09c75806077b9f5a9705987692"} Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.678090 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.813537 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb\") pod \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.813700 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config\") pod \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.813779 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc\") pod \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.813806 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwld7\" (UniqueName: \"kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7\") pod \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.813901 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb\") pod \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\" (UID: \"e8a10d6e-f132-43fe-ac06-cf1a99c9146f\") " Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.823016 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7" (OuterVolumeSpecName: "kube-api-access-qwld7") pod "e8a10d6e-f132-43fe-ac06-cf1a99c9146f" (UID: "e8a10d6e-f132-43fe-ac06-cf1a99c9146f"). InnerVolumeSpecName "kube-api-access-qwld7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.880305 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config" (OuterVolumeSpecName: "config") pod "e8a10d6e-f132-43fe-ac06-cf1a99c9146f" (UID: "e8a10d6e-f132-43fe-ac06-cf1a99c9146f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.892555 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8a10d6e-f132-43fe-ac06-cf1a99c9146f" (UID: "e8a10d6e-f132-43fe-ac06-cf1a99c9146f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.896227 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e8a10d6e-f132-43fe-ac06-cf1a99c9146f" (UID: "e8a10d6e-f132-43fe-ac06-cf1a99c9146f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.909206 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e8a10d6e-f132-43fe-ac06-cf1a99c9146f" (UID: "e8a10d6e-f132-43fe-ac06-cf1a99c9146f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.918935 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.918959 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.918967 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.918979 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwld7\" (UniqueName: \"kubernetes.io/projected/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-kube-api-access-qwld7\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:30 crc kubenswrapper[5002]: I1014 08:08:30.918990 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e8a10d6e-f132-43fe-ac06-cf1a99c9146f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:31 crc kubenswrapper[5002]: I1014 08:08:31.634570 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd85975c7-6cwfb" Oct 14 08:08:31 crc kubenswrapper[5002]: I1014 08:08:31.676309 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:08:31 crc kubenswrapper[5002]: I1014 08:08:31.683317 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bd85975c7-6cwfb"] Oct 14 08:08:31 crc kubenswrapper[5002]: I1014 08:08:31.732134 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" path="/var/lib/kubelet/pods/e8a10d6e-f132-43fe-ac06-cf1a99c9146f/volumes" Oct 14 08:08:31 crc kubenswrapper[5002]: I1014 08:08:31.989278 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:32 crc kubenswrapper[5002]: I1014 08:08:32.078714 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:33 crc kubenswrapper[5002]: I1014 08:08:33.659101 5002 generic.go:334] "Generic (PLEG): container finished" podID="0307318a-577d-4b25-9972-274b46d62f36" containerID="c4b73f9c2443ba98ce446154d4c87a5d39eb6f2bfb951754a550de379f4c2a09" exitCode=0 Oct 14 08:08:33 crc kubenswrapper[5002]: I1014 08:08:33.659327 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-v746c" event={"ID":"0307318a-577d-4b25-9972-274b46d62f36","Type":"ContainerDied","Data":"c4b73f9c2443ba98ce446154d4c87a5d39eb6f2bfb951754a550de379f4c2a09"} Oct 14 08:08:33 crc kubenswrapper[5002]: I1014 08:08:33.878112 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.009916 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-67d6789ffb-z5pvg" Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.060527 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.060779 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" containerID="cri-o://602a385670a710c0f2a0160b6dc7a2325c485a720d8642bba52663af13f990aa" gracePeriod=30 Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.060932 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" containerID="cri-o://5cab479b92c0dc5af912937d80d978e5dd7c1f03d9435b3e06270666b537f991" gracePeriod=30 Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.688999 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerID="602a385670a710c0f2a0160b6dc7a2325c485a720d8642bba52663af13f990aa" exitCode=143 Oct 14 08:08:34 crc kubenswrapper[5002]: I1014 08:08:34.689066 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerDied","Data":"602a385670a710c0f2a0160b6dc7a2325c485a720d8642bba52663af13f990aa"} Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.226741 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:36498->10.217.0.154:9311: read: connection reset by peer" Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.226874 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:36484->10.217.0.154:9311: read: connection reset by peer" Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.715984 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerID="5cab479b92c0dc5af912937d80d978e5dd7c1f03d9435b3e06270666b537f991" exitCode=0 Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.716107 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerDied","Data":"5cab479b92c0dc5af912937d80d978e5dd7c1f03d9435b3e06270666b537f991"} Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.944724 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.946814 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-central-agent" containerID="cri-o://124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e" gracePeriod=30 Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.946975 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="proxy-httpd" containerID="cri-o://6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6" gracePeriod=30 Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.947059 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-notification-agent" containerID="cri-o://cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe" gracePeriod=30 Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.947188 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="sg-core" containerID="cri-o://7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466" gracePeriod=30 Oct 14 08:08:37 crc kubenswrapper[5002]: I1014 08:08:37.975020 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.149:3000/\": read tcp 10.217.0.2:54204->10.217.0.149:3000: read: connection reset by peer" Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726760 5002 generic.go:334] "Generic (PLEG): container finished" podID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerID="6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6" exitCode=0 Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726804 5002 generic.go:334] "Generic (PLEG): container finished" podID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerID="7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466" exitCode=2 Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726813 5002 generic.go:334] "Generic (PLEG): container finished" podID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerID="124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e" exitCode=0 Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726832 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerDied","Data":"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6"} Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726942 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerDied","Data":"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466"} Oct 14 08:08:38 crc kubenswrapper[5002]: I1014 08:08:38.726953 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerDied","Data":"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e"} Oct 14 08:08:39 crc kubenswrapper[5002]: I1014 08:08:39.218352 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:08:39 crc kubenswrapper[5002]: I1014 08:08:39.218705 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:08:39 crc kubenswrapper[5002]: I1014 08:08:39.902539 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b7cd67f54-t9dgw" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.140:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.140:8443: connect: connection refused" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.140678 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-v746c" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.226955 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.227469 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtj4l\" (UniqueName: \"kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.227524 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.227763 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.227799 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.227892 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id\") pod \"0307318a-577d-4b25-9972-274b46d62f36\" (UID: \"0307318a-577d-4b25-9972-274b46d62f36\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.228057 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.228529 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0307318a-577d-4b25-9972-274b46d62f36-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.274158 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l" (OuterVolumeSpecName: "kube-api-access-mtj4l") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "kube-api-access-mtj4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.275637 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.279616 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts" (OuterVolumeSpecName: "scripts") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.330499 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtj4l\" (UniqueName: \"kubernetes.io/projected/0307318a-577d-4b25-9972-274b46d62f36-kube-api-access-mtj4l\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.330536 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.330548 5002 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.367379 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data" (OuterVolumeSpecName: "config-data") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.374038 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0307318a-577d-4b25-9972-274b46d62f36" (UID: "0307318a-577d-4b25-9972-274b46d62f36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.396185 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.440477 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs\") pod \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.440543 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle\") pod \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.440610 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qwlp\" (UniqueName: \"kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp\") pod \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.440731 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom\") pod \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.440760 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data\") pod \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\" (UID: \"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076\") " Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.441142 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.441158 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307318a-577d-4b25-9972-274b46d62f36-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.441193 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs" (OuterVolumeSpecName: "logs") pod "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" (UID: "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.444730 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" (UID: "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.445425 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp" (OuterVolumeSpecName: "kube-api-access-6qwlp") pod "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" (UID: "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076"). InnerVolumeSpecName "kube-api-access-6qwlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.465972 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" (UID: "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.490388 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data" (OuterVolumeSpecName: "config-data") pod "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" (UID: "f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.542239 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qwlp\" (UniqueName: \"kubernetes.io/projected/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-kube-api-access-6qwlp\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.542272 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.542281 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.542289 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.542298 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.748055 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74b55bb47b-hbxr2" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.748048 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74b55bb47b-hbxr2" event={"ID":"f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076","Type":"ContainerDied","Data":"82e1813334b0e56d814271a9fee5947a00216ee43737c89871617d81e63c33c3"} Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.748172 5002 scope.go:117] "RemoveContainer" containerID="5cab479b92c0dc5af912937d80d978e5dd7c1f03d9435b3e06270666b537f991" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.753154 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-v746c" event={"ID":"0307318a-577d-4b25-9972-274b46d62f36","Type":"ContainerDied","Data":"79b412ee52c9513c5c3a15c3cb804a430281432665ad30341481316e6456ea88"} Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.753188 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b412ee52c9513c5c3a15c3cb804a430281432665ad30341481316e6456ea88" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.753190 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-v746c" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.755355 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jv98l" event={"ID":"fce7a28b-9ac5-4a32-a9c9-859df10c79c0","Type":"ContainerStarted","Data":"05979b86aff0e92996725e7f20034b603345a38e9a4d81b39d9a0d60164377cf"} Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.766565 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c0df2486-02ae-48b5-b3b9-acb14ec03a6f","Type":"ContainerStarted","Data":"5a87244c8bf4c728b327d9a1701f186ca02205067c10bdb248cf87bc03ec2152"} Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.775071 5002 scope.go:117] "RemoveContainer" containerID="602a385670a710c0f2a0160b6dc7a2325c485a720d8642bba52663af13f990aa" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.779312 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jv98l" podStartSLOduration=1.647800299 podStartE2EDuration="1m13.779301897s" podCreationTimestamp="2025-10-14 08:07:27 +0000 UTC" firstStartedPulling="2025-10-14 08:07:28.012016348 +0000 UTC m=+980.993255800" lastFinishedPulling="2025-10-14 08:08:40.143517946 +0000 UTC m=+1053.124757398" observedRunningTime="2025-10-14 08:08:40.776408729 +0000 UTC m=+1053.757648201" watchObservedRunningTime="2025-10-14 08:08:40.779301897 +0000 UTC m=+1053.760541349" Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.803190 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.809399 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-74b55bb47b-hbxr2"] Oct 14 08:08:40 crc kubenswrapper[5002]: I1014 08:08:40.818872 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.7998185100000001 podStartE2EDuration="11.818837347s" podCreationTimestamp="2025-10-14 08:08:29 +0000 UTC" firstStartedPulling="2025-10-14 08:08:30.15290619 +0000 UTC m=+1043.134145642" lastFinishedPulling="2025-10-14 08:08:40.171924987 +0000 UTC m=+1053.153164479" observedRunningTime="2025-10-14 08:08:40.810921035 +0000 UTC m=+1053.792160487" watchObservedRunningTime="2025-10-14 08:08:40.818837347 +0000 UTC m=+1053.800076799" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468290 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:41 crc kubenswrapper[5002]: E1014 08:08:41.468858 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468871 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" Oct 14 08:08:41 crc kubenswrapper[5002]: E1014 08:08:41.468884 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0307318a-577d-4b25-9972-274b46d62f36" containerName="cinder-db-sync" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468890 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0307318a-577d-4b25-9972-274b46d62f36" containerName="cinder-db-sync" Oct 14 08:08:41 crc kubenswrapper[5002]: E1014 08:08:41.468914 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="dnsmasq-dns" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468921 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="dnsmasq-dns" Oct 14 08:08:41 crc kubenswrapper[5002]: E1014 08:08:41.468928 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468934 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" Oct 14 08:08:41 crc kubenswrapper[5002]: E1014 08:08:41.468945 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="init" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.468951 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="init" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.469102 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0307318a-577d-4b25-9972-274b46d62f36" containerName="cinder-db-sync" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.469119 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8a10d6e-f132-43fe-ac06-cf1a99c9146f" containerName="dnsmasq-dns" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.474938 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.475026 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.476235 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.479393 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.479699 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.479926 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.480720 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-xbc8s" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.484246 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.540120 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.541681 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.551963 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581013 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581178 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581269 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581306 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581351 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581539 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhcl6\" (UniqueName: \"kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581600 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8c8p\" (UniqueName: \"kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581643 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581765 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.581962 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.582047 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.680419 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.681741 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683217 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683257 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683276 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683307 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683346 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhcl6\" (UniqueName: \"kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683367 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8c8p\" (UniqueName: \"kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683384 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683410 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683436 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683464 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.683495 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.684385 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.685610 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.692454 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.693864 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.694013 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.694611 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.695346 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.698192 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.698187 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.703067 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.710223 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.712683 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8c8p\" (UniqueName: \"kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p\") pod \"dnsmasq-dns-68fcd76675-6nqbq\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.722433 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhcl6\" (UniqueName: \"kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6\") pod \"cinder-scheduler-0\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.732822 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" path="/var/lib/kubelet/pods/f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076/volumes" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785128 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785492 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785509 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785554 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785573 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785591 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfd7c\" (UniqueName: \"kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.785612 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.793874 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.872523 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.889124 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.889250 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.889980 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.889588 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.890149 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.890181 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.890203 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfd7c\" (UniqueName: \"kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.890248 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.890721 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.895239 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.896224 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.898585 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.906669 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfd7c\" (UniqueName: \"kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:41 crc kubenswrapper[5002]: I1014 08:08:41.910225 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts\") pod \"cinder-api-0\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " pod="openstack/cinder-api-0" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.084992 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.270075 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.434611 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.572480 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.687941 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:42 crc kubenswrapper[5002]: W1014 08:08:42.697380 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc838caa2_0c0e_45b2_b1d4_85b04895f1ca.slice/crio-dcc8d8dd3227e42e1c233dcb92c4fc6221cbdee4ea4ec4502bb1ea969a388744 WatchSource:0}: Error finding container dcc8d8dd3227e42e1c233dcb92c4fc6221cbdee4ea4ec4502bb1ea969a388744: Status 404 returned error can't find the container with id dcc8d8dd3227e42e1c233dcb92c4fc6221cbdee4ea4ec4502bb1ea969a388744 Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767398 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767497 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5gtv\" (UniqueName: \"kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767533 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767610 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767652 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767674 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767691 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data\") pod \"866e36d1-d1ee-49c8-8052-fa5097ca695b\" (UID: \"866e36d1-d1ee-49c8-8052-fa5097ca695b\") " Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.767898 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.768130 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.768018 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.772720 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv" (OuterVolumeSpecName: "kube-api-access-m5gtv") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "kube-api-access-m5gtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.773353 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts" (OuterVolumeSpecName: "scripts") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.793010 5002 generic.go:334] "Generic (PLEG): container finished" podID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerID="0de5d4de7980754fcab606ab40622637524494c1bbc322adeb7ade3b688bd60d" exitCode=0 Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.793083 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" event={"ID":"6946ff35-58ec-4de0-be57-f077bbb50c63","Type":"ContainerDied","Data":"0de5d4de7980754fcab606ab40622637524494c1bbc322adeb7ade3b688bd60d"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.793157 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" event={"ID":"6946ff35-58ec-4de0-be57-f077bbb50c63","Type":"ContainerStarted","Data":"d8d43b40b822bb4c7e60152ae106293ecd216e826a7b5c1daf6601d9af7bb8d6"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.794139 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerStarted","Data":"d82c5946f200b3a7a7e0a5b41b5b0f6ae8192b984959e127f6eeda9c4216c9fa"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.798849 5002 generic.go:334] "Generic (PLEG): container finished" podID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerID="cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe" exitCode=0 Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.798983 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerDied","Data":"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.799075 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"866e36d1-d1ee-49c8-8052-fa5097ca695b","Type":"ContainerDied","Data":"d8c8bc26c7664de28159988f5e4f6b038f9f4671ab88a72f0099407690a419bd"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.799138 5002 scope.go:117] "RemoveContainer" containerID="6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.799322 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.801019 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.810031 5002 generic.go:334] "Generic (PLEG): container finished" podID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" containerID="05979b86aff0e92996725e7f20034b603345a38e9a4d81b39d9a0d60164377cf" exitCode=0 Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.810101 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jv98l" event={"ID":"fce7a28b-9ac5-4a32-a9c9-859df10c79c0","Type":"ContainerDied","Data":"05979b86aff0e92996725e7f20034b603345a38e9a4d81b39d9a0d60164377cf"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.811208 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerStarted","Data":"dcc8d8dd3227e42e1c233dcb92c4fc6221cbdee4ea4ec4502bb1ea969a388744"} Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.859315 5002 scope.go:117] "RemoveContainer" containerID="7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.873122 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.873373 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5gtv\" (UniqueName: \"kubernetes.io/projected/866e36d1-d1ee-49c8-8052-fa5097ca695b-kube-api-access-m5gtv\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.873389 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.873398 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/866e36d1-d1ee-49c8-8052-fa5097ca695b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.874781 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data" (OuterVolumeSpecName: "config-data") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.883126 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "866e36d1-d1ee-49c8-8052-fa5097ca695b" (UID: "866e36d1-d1ee-49c8-8052-fa5097ca695b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.899786 5002 scope.go:117] "RemoveContainer" containerID="cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.920181 5002 scope.go:117] "RemoveContainer" containerID="124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.964969 5002 scope.go:117] "RemoveContainer" containerID="6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6" Oct 14 08:08:42 crc kubenswrapper[5002]: E1014 08:08:42.965398 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6\": container with ID starting with 6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6 not found: ID does not exist" containerID="6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.965488 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6"} err="failed to get container status \"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6\": rpc error: code = NotFound desc = could not find container \"6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6\": container with ID starting with 6bfa982da3f4d925e660d199db4ed08bd00ffa38fb35e31ab0c47e954fbfa1d6 not found: ID does not exist" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.965515 5002 scope.go:117] "RemoveContainer" containerID="7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466" Oct 14 08:08:42 crc kubenswrapper[5002]: E1014 08:08:42.965921 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466\": container with ID starting with 7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466 not found: ID does not exist" containerID="7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.965972 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466"} err="failed to get container status \"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466\": rpc error: code = NotFound desc = could not find container \"7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466\": container with ID starting with 7e154a2981672ebd6aea857fddafb881d2ee628b0806950142a0211ff7dee466 not found: ID does not exist" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.966005 5002 scope.go:117] "RemoveContainer" containerID="cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe" Oct 14 08:08:42 crc kubenswrapper[5002]: E1014 08:08:42.966407 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe\": container with ID starting with cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe not found: ID does not exist" containerID="cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.966458 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe"} err="failed to get container status \"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe\": rpc error: code = NotFound desc = could not find container \"cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe\": container with ID starting with cbac0cb55c4cea3b9578437da185c82b9543808292339e06effa1d2ac0ac8afe not found: ID does not exist" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.966473 5002 scope.go:117] "RemoveContainer" containerID="124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e" Oct 14 08:08:42 crc kubenswrapper[5002]: E1014 08:08:42.966726 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e\": container with ID starting with 124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e not found: ID does not exist" containerID="124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.966755 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e"} err="failed to get container status \"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e\": rpc error: code = NotFound desc = could not find container \"124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e\": container with ID starting with 124fb1be2a72af37a401bb7dc8f69bd34d7a004f25d889658da58b5ccbf4660e not found: ID does not exist" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.975243 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:42 crc kubenswrapper[5002]: I1014 08:08:42.975281 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866e36d1-d1ee-49c8-8052-fa5097ca695b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.138061 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.146481 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179176 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:43 crc kubenswrapper[5002]: E1014 08:08:43.179547 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-notification-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179558 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-notification-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: E1014 08:08:43.179573 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-central-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179581 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-central-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: E1014 08:08:43.179599 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="proxy-httpd" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179606 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="proxy-httpd" Oct 14 08:08:43 crc kubenswrapper[5002]: E1014 08:08:43.179617 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="sg-core" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179623 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="sg-core" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179768 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-central-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179781 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="proxy-httpd" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179795 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="sg-core" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.179805 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" containerName="ceilometer-notification-agent" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.181303 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.186226 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.186363 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.207550 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281099 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281150 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281204 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281245 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281335 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281380 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqs9l\" (UniqueName: \"kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.281452 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382737 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382800 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382857 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382902 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382934 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.382960 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqs9l\" (UniqueName: \"kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.383009 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.383749 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.383931 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.389623 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.390472 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.391176 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.399060 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.400671 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.410446 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqs9l\" (UniqueName: \"kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l\") pod \"ceilometer-0\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.506516 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.730228 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866e36d1-d1ee-49c8-8052-fa5097ca695b" path="/var/lib/kubelet/pods/866e36d1-d1ee-49c8-8052-fa5097ca695b/volumes" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.829526 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" event={"ID":"6946ff35-58ec-4de0-be57-f077bbb50c63","Type":"ContainerStarted","Data":"422d62c0af860c269d29d2f2a479c116db633b500492cd3c298a01ec8cdc3e89"} Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.829781 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.843761 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerStarted","Data":"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03"} Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.848750 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" podStartSLOduration=2.8487343750000003 podStartE2EDuration="2.848734375s" podCreationTimestamp="2025-10-14 08:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:43.846813704 +0000 UTC m=+1056.828053166" watchObservedRunningTime="2025-10-14 08:08:43.848734375 +0000 UTC m=+1056.829973817" Oct 14 08:08:43 crc kubenswrapper[5002]: I1014 08:08:43.981237 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.277682 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jv98l" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.309772 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs\") pod \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.309863 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sk7j6\" (UniqueName: \"kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6\") pod \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.309887 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts\") pod \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.309973 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data\") pod \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.310012 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle\") pod \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\" (UID: \"fce7a28b-9ac5-4a32-a9c9-859df10c79c0\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.311452 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs" (OuterVolumeSpecName: "logs") pod "fce7a28b-9ac5-4a32-a9c9-859df10c79c0" (UID: "fce7a28b-9ac5-4a32-a9c9-859df10c79c0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.335949 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts" (OuterVolumeSpecName: "scripts") pod "fce7a28b-9ac5-4a32-a9c9-859df10c79c0" (UID: "fce7a28b-9ac5-4a32-a9c9-859df10c79c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.361116 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fce7a28b-9ac5-4a32-a9c9-859df10c79c0" (UID: "fce7a28b-9ac5-4a32-a9c9-859df10c79c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.404065 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6" (OuterVolumeSpecName: "kube-api-access-sk7j6") pod "fce7a28b-9ac5-4a32-a9c9-859df10c79c0" (UID: "fce7a28b-9ac5-4a32-a9c9-859df10c79c0"). InnerVolumeSpecName "kube-api-access-sk7j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.420896 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.420933 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sk7j6\" (UniqueName: \"kubernetes.io/projected/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-kube-api-access-sk7j6\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.420944 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.420953 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.424134 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data" (OuterVolumeSpecName: "config-data") pod "fce7a28b-9ac5-4a32-a9c9-859df10c79c0" (UID: "fce7a28b-9ac5-4a32-a9c9-859df10c79c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.522739 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce7a28b-9ac5-4a32-a9c9-859df10c79c0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.589809 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624318 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624458 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624595 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44nqs\" (UniqueName: \"kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624679 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624747 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624777 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs" (OuterVolumeSpecName: "logs") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624783 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.624892 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key\") pod \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\" (UID: \"9085e7e2-47b0-4121-bb58-3c633ab7e00d\") " Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.625228 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9085e7e2-47b0-4121-bb58-3c633ab7e00d-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.634378 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.640120 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs" (OuterVolumeSpecName: "kube-api-access-44nqs") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "kube-api-access-44nqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.666294 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts" (OuterVolumeSpecName: "scripts") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.669777 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.691437 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data" (OuterVolumeSpecName: "config-data") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.701618 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "9085e7e2-47b0-4121-bb58-3c633ab7e00d" (UID: "9085e7e2-47b0-4121-bb58-3c633ab7e00d"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726324 5002 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726364 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726373 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9085e7e2-47b0-4121-bb58-3c633ab7e00d-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726382 5002 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726391 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9085e7e2-47b0-4121-bb58-3c633ab7e00d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.726399 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44nqs\" (UniqueName: \"kubernetes.io/projected/9085e7e2-47b0-4121-bb58-3c633ab7e00d-kube-api-access-44nqs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.862340 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jv98l" event={"ID":"fce7a28b-9ac5-4a32-a9c9-859df10c79c0","Type":"ContainerDied","Data":"1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.862387 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1549cf1ec7170d14301ca21780c8e234960743dd75b63cac925ce7d87b47e0ff" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.862399 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jv98l" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.871349 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerStarted","Data":"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.872962 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api-log" containerID="cri-o://c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" gracePeriod=30 Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.873675 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.874993 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api" containerID="cri-o://75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" gracePeriod=30 Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.889472 5002 generic.go:334] "Generic (PLEG): container finished" podID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerID="ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990" exitCode=137 Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.889605 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b7cd67f54-t9dgw" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.889608 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerDied","Data":"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.889674 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b7cd67f54-t9dgw" event={"ID":"9085e7e2-47b0-4121-bb58-3c633ab7e00d","Type":"ContainerDied","Data":"968cf559e1c3aca3a2260b46db5cf82aaab1cc5e2cd652c788e13575bf224f56"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.889693 5002 scope.go:117] "RemoveContainer" containerID="b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.900824 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerStarted","Data":"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.900874 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerStarted","Data":"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.910082 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerStarted","Data":"01319d05524c3663aa130456f9f276c8bb2abc128cad32efd381546ea91b6e9f"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.910112 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerStarted","Data":"061023af8273d2e046acb34e23d26987c102a9695d9ef8a1da26e63b05620a83"} Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.919328 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.919307286 podStartE2EDuration="3.919307286s" podCreationTimestamp="2025-10-14 08:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:44.893108614 +0000 UTC m=+1057.874348076" watchObservedRunningTime="2025-10-14 08:08:44.919307286 +0000 UTC m=+1057.900546738" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.937531 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.957857921 podStartE2EDuration="3.937510845s" podCreationTimestamp="2025-10-14 08:08:41 +0000 UTC" firstStartedPulling="2025-10-14 08:08:42.260564142 +0000 UTC m=+1055.241803584" lastFinishedPulling="2025-10-14 08:08:43.240217056 +0000 UTC m=+1056.221456508" observedRunningTime="2025-10-14 08:08:44.92167974 +0000 UTC m=+1057.902919202" watchObservedRunningTime="2025-10-14 08:08:44.937510845 +0000 UTC m=+1057.918750297" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.954280 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.971481 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b7cd67f54-t9dgw"] Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.989898 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5dc947dd58-kwj5n"] Oct 14 08:08:44 crc kubenswrapper[5002]: E1014 08:08:44.990327 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990344 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" Oct 14 08:08:44 crc kubenswrapper[5002]: E1014 08:08:44.990369 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" containerName="placement-db-sync" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990376 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" containerName="placement-db-sync" Oct 14 08:08:44 crc kubenswrapper[5002]: E1014 08:08:44.990397 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon-log" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990402 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon-log" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990593 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990616 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" containerName="horizon-log" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.990624 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" containerName="placement-db-sync" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.991558 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.994006 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.994251 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.994414 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.994731 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gvlg9" Oct 14 08:08:44 crc kubenswrapper[5002]: I1014 08:08:44.994811 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.022677 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5dc947dd58-kwj5n"] Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.031978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-public-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032073 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-combined-ca-bundle\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032102 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-scripts\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032127 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-internal-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032145 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0141dd41-3526-47f8-8477-f0d347c58cea-logs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032178 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-config-data\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.032221 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n88f\" (UniqueName: \"kubernetes.io/projected/0141dd41-3526-47f8-8477-f0d347c58cea-kube-api-access-7n88f\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.110754 5002 scope.go:117] "RemoveContainer" containerID="ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.124560 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc838caa2_0c0e_45b2_b1d4_85b04895f1ca.slice/crio-c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfce7a28b_9ac5_4a32_a9c9_859df10c79c0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9085e7e2_47b0_4121_bb58_3c633ab7e00d.slice/crio-968cf559e1c3aca3a2260b46db5cf82aaab1cc5e2cd652c788e13575bf224f56\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc838caa2_0c0e_45b2_b1d4_85b04895f1ca.slice/crio-conmon-c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.126749 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.127176 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-74b55bb47b-hbxr2" podUID="f4b0f4f9-d9f8-4675-b8a9-0b4bf64e4076" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.134971 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n88f\" (UniqueName: \"kubernetes.io/projected/0141dd41-3526-47f8-8477-f0d347c58cea-kube-api-access-7n88f\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135016 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-public-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135091 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-combined-ca-bundle\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135120 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-scripts\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135144 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-internal-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135166 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0141dd41-3526-47f8-8477-f0d347c58cea-logs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.135197 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-config-data\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.138930 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0141dd41-3526-47f8-8477-f0d347c58cea-logs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.143950 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-internal-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.144031 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-combined-ca-bundle\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.145751 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-public-tls-certs\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.146083 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-scripts\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.147325 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0141dd41-3526-47f8-8477-f0d347c58cea-config-data\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.168436 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n88f\" (UniqueName: \"kubernetes.io/projected/0141dd41-3526-47f8-8477-f0d347c58cea-kube-api-access-7n88f\") pod \"placement-5dc947dd58-kwj5n\" (UID: \"0141dd41-3526-47f8-8477-f0d347c58cea\") " pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.318124 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.323862 5002 scope.go:117] "RemoveContainer" containerID="b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.324281 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2\": container with ID starting with b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2 not found: ID does not exist" containerID="b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.324313 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2"} err="failed to get container status \"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2\": rpc error: code = NotFound desc = could not find container \"b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2\": container with ID starting with b961b8b960ba1d86da4e66f2b8a986633688ab8d7896cb3b3c4f264b9bea5fd2 not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.324332 5002 scope.go:117] "RemoveContainer" containerID="ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.324619 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990\": container with ID starting with ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990 not found: ID does not exist" containerID="ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.324637 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990"} err="failed to get container status \"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990\": rpc error: code = NotFound desc = could not find container \"ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990\": container with ID starting with ea006ce0d4f36ec2dd3af5404e3e8e22661455105f08b4a553d42f6a4c845990 not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.502991 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541156 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541214 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfd7c\" (UniqueName: \"kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541249 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541265 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541336 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541359 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541379 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541400 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom\") pod \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\" (UID: \"c838caa2-0c0e-45b2-b1d4-85b04895f1ca\") " Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541668 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs" (OuterVolumeSpecName: "logs") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541776 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.541789 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.550989 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts" (OuterVolumeSpecName: "scripts") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.551011 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.551050 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c" (OuterVolumeSpecName: "kube-api-access-rfd7c") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "kube-api-access-rfd7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.576080 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.590469 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data" (OuterVolumeSpecName: "config-data") pod "c838caa2-0c0e-45b2-b1d4-85b04895f1ca" (UID: "c838caa2-0c0e-45b2-b1d4-85b04895f1ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.643419 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.643460 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.643473 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.643487 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.643499 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfd7c\" (UniqueName: \"kubernetes.io/projected/c838caa2-0c0e-45b2-b1d4-85b04895f1ca-kube-api-access-rfd7c\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.647970 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.762813 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9085e7e2-47b0-4121-bb58-3c633ab7e00d" path="/var/lib/kubelet/pods/9085e7e2-47b0-4121-bb58-3c633ab7e00d/volumes" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.824717 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5dc947dd58-kwj5n"] Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.920445 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5dc947dd58-kwj5n" event={"ID":"0141dd41-3526-47f8-8477-f0d347c58cea","Type":"ContainerStarted","Data":"9d80af67e61931a2444ef5b640e8e1b84ace6f056e59c8b367b5df8b172c324c"} Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.923366 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerStarted","Data":"b636e5a146d90ec398735b2f28bffc2bdad71f267f9ba156161af9ee86493944"} Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926297 5002 generic.go:334] "Generic (PLEG): container finished" podID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerID="75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" exitCode=0 Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926328 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926354 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerDied","Data":"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c"} Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926387 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerDied","Data":"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03"} Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926407 5002 scope.go:117] "RemoveContainer" containerID="75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926337 5002 generic.go:334] "Generic (PLEG): container finished" podID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerID="c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" exitCode=143 Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.926524 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c838caa2-0c0e-45b2-b1d4-85b04895f1ca","Type":"ContainerDied","Data":"dcc8d8dd3227e42e1c233dcb92c4fc6221cbdee4ea4ec4502bb1ea969a388744"} Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.947071 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.953788 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.956004 5002 scope.go:117] "RemoveContainer" containerID="c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.975239 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.976710 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.976821 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.976959 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api-log" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.976976 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api-log" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.977849 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.977886 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" containerName="cinder-api-log" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.978998 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.984486 5002 scope.go:117] "RemoveContainer" containerID="75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.984911 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.985121 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.985357 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.985555 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c\": container with ID starting with 75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c not found: ID does not exist" containerID="75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.985583 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c"} err="failed to get container status \"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c\": rpc error: code = NotFound desc = could not find container \"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c\": container with ID starting with 75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.985603 5002 scope.go:117] "RemoveContainer" containerID="c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" Oct 14 08:08:45 crc kubenswrapper[5002]: E1014 08:08:45.986705 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03\": container with ID starting with c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03 not found: ID does not exist" containerID="c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.986728 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03"} err="failed to get container status \"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03\": rpc error: code = NotFound desc = could not find container \"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03\": container with ID starting with c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03 not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.986743 5002 scope.go:117] "RemoveContainer" containerID="75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.987183 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c"} err="failed to get container status \"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c\": rpc error: code = NotFound desc = could not find container \"75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c\": container with ID starting with 75f8ea1b177cc258f8a8e22121eb1e04c1b26070fcba4a04a67a3c3b3c97d26c not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.987202 5002 scope.go:117] "RemoveContainer" containerID="c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.989479 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03"} err="failed to get container status \"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03\": rpc error: code = NotFound desc = could not find container \"c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03\": container with ID starting with c49908cfe9b2907e7efd22033e031ed49a5634d378ff6257a0f95a75ec0a1a03 not found: ID does not exist" Oct 14 08:08:45 crc kubenswrapper[5002]: I1014 08:08:45.990872 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051349 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29ae9ac-13f1-4621-9696-4eabea39109f-logs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051441 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051469 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051500 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-scripts\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051521 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051541 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29ae9ac-13f1-4621-9696-4eabea39109f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051588 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051606 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.051644 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsm97\" (UniqueName: \"kubernetes.io/projected/d29ae9ac-13f1-4621-9696-4eabea39109f-kube-api-access-dsm97\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.152863 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29ae9ac-13f1-4621-9696-4eabea39109f-logs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.152921 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.152945 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.152974 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-scripts\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.152996 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.153016 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29ae9ac-13f1-4621-9696-4eabea39109f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.153061 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.153083 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.153122 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsm97\" (UniqueName: \"kubernetes.io/projected/d29ae9ac-13f1-4621-9696-4eabea39109f-kube-api-access-dsm97\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.155377 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29ae9ac-13f1-4621-9696-4eabea39109f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.157264 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.157613 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29ae9ac-13f1-4621-9696-4eabea39109f-logs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.162981 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.165533 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.166069 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-config-data\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.166379 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-scripts\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.167068 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d29ae9ac-13f1-4621-9696-4eabea39109f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.176755 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsm97\" (UniqueName: \"kubernetes.io/projected/d29ae9ac-13f1-4621-9696-4eabea39109f-kube-api-access-dsm97\") pod \"cinder-api-0\" (UID: \"d29ae9ac-13f1-4621-9696-4eabea39109f\") " pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.302309 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 08:08:46 crc kubenswrapper[5002]: W1014 08:08:46.771554 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29ae9ac_13f1_4621_9696_4eabea39109f.slice/crio-4a1b04e00016466a488b4fe23bf83f0f035de6cb78465381c81ba6be821cb8e8 WatchSource:0}: Error finding container 4a1b04e00016466a488b4fe23bf83f0f035de6cb78465381c81ba6be821cb8e8: Status 404 returned error can't find the container with id 4a1b04e00016466a488b4fe23bf83f0f035de6cb78465381c81ba6be821cb8e8 Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.772410 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.795152 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.945527 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5dc947dd58-kwj5n" event={"ID":"0141dd41-3526-47f8-8477-f0d347c58cea","Type":"ContainerStarted","Data":"36bd36c41593d03582f25c36d19b7d636c607057851b0dd73ca326f44025ea8c"} Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.948255 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5dc947dd58-kwj5n" event={"ID":"0141dd41-3526-47f8-8477-f0d347c58cea","Type":"ContainerStarted","Data":"7db5d9ae4b3e82c209f94ad9d562344d7e00496803feaffe685f40d25ecbf3ee"} Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.948290 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.948302 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.949153 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29ae9ac-13f1-4621-9696-4eabea39109f","Type":"ContainerStarted","Data":"4a1b04e00016466a488b4fe23bf83f0f035de6cb78465381c81ba6be821cb8e8"} Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.952746 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerStarted","Data":"6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892"} Oct 14 08:08:46 crc kubenswrapper[5002]: I1014 08:08:46.972936 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5dc947dd58-kwj5n" podStartSLOduration=2.972920372 podStartE2EDuration="2.972920372s" podCreationTimestamp="2025-10-14 08:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:46.963122869 +0000 UTC m=+1059.944362321" watchObservedRunningTime="2025-10-14 08:08:46.972920372 +0000 UTC m=+1059.954159824" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.391229 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4ks2z"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.392357 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.403046 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4ks2z"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.484980 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j58n7\" (UniqueName: \"kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7\") pod \"nova-api-db-create-4ks2z\" (UID: \"02a2af3f-d7e5-4e1e-89ba-b567ab162985\") " pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.507905 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-575kw"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.513091 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.523190 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-575kw"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.586628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j58n7\" (UniqueName: \"kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7\") pod \"nova-api-db-create-4ks2z\" (UID: \"02a2af3f-d7e5-4e1e-89ba-b567ab162985\") " pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.614176 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j58n7\" (UniqueName: \"kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7\") pod \"nova-api-db-create-4ks2z\" (UID: \"02a2af3f-d7e5-4e1e-89ba-b567ab162985\") " pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.692994 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmvwp\" (UniqueName: \"kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp\") pod \"nova-cell0-db-create-575kw\" (UID: \"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8\") " pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.707116 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-c2mx5"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.708257 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.721207 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c2mx5"] Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.759381 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.774905 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c838caa2-0c0e-45b2-b1d4-85b04895f1ca" path="/var/lib/kubelet/pods/c838caa2-0c0e-45b2-b1d4-85b04895f1ca/volumes" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.794901 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26zk6\" (UniqueName: \"kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6\") pod \"nova-cell1-db-create-c2mx5\" (UID: \"45065e0e-2e90-427d-a49c-423e2ba340d7\") " pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.795024 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmvwp\" (UniqueName: \"kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp\") pod \"nova-cell0-db-create-575kw\" (UID: \"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8\") " pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.814241 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmvwp\" (UniqueName: \"kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp\") pod \"nova-cell0-db-create-575kw\" (UID: \"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8\") " pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.896257 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26zk6\" (UniqueName: \"kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6\") pod \"nova-cell1-db-create-c2mx5\" (UID: \"45065e0e-2e90-427d-a49c-423e2ba340d7\") " pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.919159 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26zk6\" (UniqueName: \"kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6\") pod \"nova-cell1-db-create-c2mx5\" (UID: \"45065e0e-2e90-427d-a49c-423e2ba340d7\") " pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.964434 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29ae9ac-13f1-4621-9696-4eabea39109f","Type":"ContainerStarted","Data":"0e7a5fb622212684cdeeb397421ef1b22146a3a0a9acba2af7dd23207caebbac"} Oct 14 08:08:47 crc kubenswrapper[5002]: I1014 08:08:47.982660 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.027687 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.139287 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7855f7b865-dtwlh" Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.189335 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.200603 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cfc656d-2j56s" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-api" containerID="cri-o://9e71e0cf8fbe795bdbcf8d98fcb033b339f5649ea1fe370d253c12cd71a8129b" gracePeriod=30 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.201020 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7cfc656d-2j56s" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-httpd" containerID="cri-o://69ae80016f8c5320fc50b1b9bce8c35caa9dcd2528dcd0d705d082389b37e845" gracePeriod=30 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.231235 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4ks2z"] Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.483072 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-575kw"] Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.644200 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-c2mx5"] Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.973332 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29ae9ac-13f1-4621-9696-4eabea39109f","Type":"ContainerStarted","Data":"942665f71045b5f02d46af369c6bd9bc6431bce88523def490284fff7626b45e"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.974450 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.976542 5002 generic.go:334] "Generic (PLEG): container finished" podID="45065e0e-2e90-427d-a49c-423e2ba340d7" containerID="51d2dd25862e928eb85bb2ca84346923b18e4a123a98b9e16d07a3adbc66de19" exitCode=0 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.976590 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c2mx5" event={"ID":"45065e0e-2e90-427d-a49c-423e2ba340d7","Type":"ContainerDied","Data":"51d2dd25862e928eb85bb2ca84346923b18e4a123a98b9e16d07a3adbc66de19"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.976608 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c2mx5" event={"ID":"45065e0e-2e90-427d-a49c-423e2ba340d7","Type":"ContainerStarted","Data":"d83daf771cfbd7e573efc8e9f225b9927b94eca06c76ea58f82149b2b6680b9d"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.979168 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerStarted","Data":"a8079d269fb4a0b25a57bc735e82c3a922f2636e9704a9f1c002012c6594dc90"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.979427 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.982367 5002 generic.go:334] "Generic (PLEG): container finished" podID="fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" containerID="66ac1cb35ce3b20dfde7f1a78effc2cea89c12fdf651b690f8f3283f733d89a2" exitCode=0 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.982423 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-575kw" event={"ID":"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8","Type":"ContainerDied","Data":"66ac1cb35ce3b20dfde7f1a78effc2cea89c12fdf651b690f8f3283f733d89a2"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.982440 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-575kw" event={"ID":"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8","Type":"ContainerStarted","Data":"52421ab1437cae6bf46de811d3ffb4e19a62721c3161f27a3f59f8dffd6c885c"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.984182 5002 generic.go:334] "Generic (PLEG): container finished" podID="02a2af3f-d7e5-4e1e-89ba-b567ab162985" containerID="1610fa5aa92070137909bf008b5d3f51f59c57b77c710f13f193a351ff1444ba" exitCode=0 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.984257 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4ks2z" event={"ID":"02a2af3f-d7e5-4e1e-89ba-b567ab162985","Type":"ContainerDied","Data":"1610fa5aa92070137909bf008b5d3f51f59c57b77c710f13f193a351ff1444ba"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.984286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4ks2z" event={"ID":"02a2af3f-d7e5-4e1e-89ba-b567ab162985","Type":"ContainerStarted","Data":"a81e49d7745611ece1fffdcd6de95e3bf0b204ac4b282274b6d76d6a4114534d"} Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.991225 5002 generic.go:334] "Generic (PLEG): container finished" podID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerID="69ae80016f8c5320fc50b1b9bce8c35caa9dcd2528dcd0d705d082389b37e845" exitCode=0 Oct 14 08:08:48 crc kubenswrapper[5002]: I1014 08:08:48.991278 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerDied","Data":"69ae80016f8c5320fc50b1b9bce8c35caa9dcd2528dcd0d705d082389b37e845"} Oct 14 08:08:49 crc kubenswrapper[5002]: I1014 08:08:49.003473 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.003456118 podStartE2EDuration="4.003456118s" podCreationTimestamp="2025-10-14 08:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:49.001536416 +0000 UTC m=+1061.982775868" watchObservedRunningTime="2025-10-14 08:08:49.003456118 +0000 UTC m=+1061.984695570" Oct 14 08:08:49 crc kubenswrapper[5002]: I1014 08:08:49.073632 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.244533689 podStartE2EDuration="6.07361689s" podCreationTimestamp="2025-10-14 08:08:43 +0000 UTC" firstStartedPulling="2025-10-14 08:08:44.021577111 +0000 UTC m=+1057.002816563" lastFinishedPulling="2025-10-14 08:08:47.850660302 +0000 UTC m=+1060.831899764" observedRunningTime="2025-10-14 08:08:49.071181714 +0000 UTC m=+1062.052421176" watchObservedRunningTime="2025-10-14 08:08:49.07361689 +0000 UTC m=+1062.054856342" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.541235 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.549144 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.554801 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.658077 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j58n7\" (UniqueName: \"kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7\") pod \"02a2af3f-d7e5-4e1e-89ba-b567ab162985\" (UID: \"02a2af3f-d7e5-4e1e-89ba-b567ab162985\") " Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.658365 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmvwp\" (UniqueName: \"kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp\") pod \"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8\" (UID: \"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8\") " Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.658543 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26zk6\" (UniqueName: \"kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6\") pod \"45065e0e-2e90-427d-a49c-423e2ba340d7\" (UID: \"45065e0e-2e90-427d-a49c-423e2ba340d7\") " Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.667737 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp" (OuterVolumeSpecName: "kube-api-access-wmvwp") pod "fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" (UID: "fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8"). InnerVolumeSpecName "kube-api-access-wmvwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.669021 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7" (OuterVolumeSpecName: "kube-api-access-j58n7") pod "02a2af3f-d7e5-4e1e-89ba-b567ab162985" (UID: "02a2af3f-d7e5-4e1e-89ba-b567ab162985"). InnerVolumeSpecName "kube-api-access-j58n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.670580 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6" (OuterVolumeSpecName: "kube-api-access-26zk6") pod "45065e0e-2e90-427d-a49c-423e2ba340d7" (UID: "45065e0e-2e90-427d-a49c-423e2ba340d7"). InnerVolumeSpecName "kube-api-access-26zk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.761272 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j58n7\" (UniqueName: \"kubernetes.io/projected/02a2af3f-d7e5-4e1e-89ba-b567ab162985-kube-api-access-j58n7\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.761522 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmvwp\" (UniqueName: \"kubernetes.io/projected/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8-kube-api-access-wmvwp\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:50 crc kubenswrapper[5002]: I1014 08:08:50.761583 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26zk6\" (UniqueName: \"kubernetes.io/projected/45065e0e-2e90-427d-a49c-423e2ba340d7-kube-api-access-26zk6\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.012990 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-c2mx5" event={"ID":"45065e0e-2e90-427d-a49c-423e2ba340d7","Type":"ContainerDied","Data":"d83daf771cfbd7e573efc8e9f225b9927b94eca06c76ea58f82149b2b6680b9d"} Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.013748 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d83daf771cfbd7e573efc8e9f225b9927b94eca06c76ea58f82149b2b6680b9d" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.013979 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-c2mx5" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.024296 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-575kw" event={"ID":"fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8","Type":"ContainerDied","Data":"52421ab1437cae6bf46de811d3ffb4e19a62721c3161f27a3f59f8dffd6c885c"} Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.024333 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52421ab1437cae6bf46de811d3ffb4e19a62721c3161f27a3f59f8dffd6c885c" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.024936 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-575kw" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.026552 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4ks2z" event={"ID":"02a2af3f-d7e5-4e1e-89ba-b567ab162985","Type":"ContainerDied","Data":"a81e49d7745611ece1fffdcd6de95e3bf0b204ac4b282274b6d76d6a4114534d"} Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.026583 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a81e49d7745611ece1fffdcd6de95e3bf0b204ac4b282274b6d76d6a4114534d" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.026632 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4ks2z" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.875109 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.974106 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:51 crc kubenswrapper[5002]: I1014 08:08:51.974333 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="dnsmasq-dns" containerID="cri-o://3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7" gracePeriod=10 Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.078604 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.148495 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.476902 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.601227 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc\") pod \"db33216f-b3e7-418f-ac05-129fb34f4538\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.601359 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwl7v\" (UniqueName: \"kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v\") pod \"db33216f-b3e7-418f-ac05-129fb34f4538\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.601391 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config\") pod \"db33216f-b3e7-418f-ac05-129fb34f4538\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.601453 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb\") pod \"db33216f-b3e7-418f-ac05-129fb34f4538\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.601515 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb\") pod \"db33216f-b3e7-418f-ac05-129fb34f4538\" (UID: \"db33216f-b3e7-418f-ac05-129fb34f4538\") " Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.625334 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v" (OuterVolumeSpecName: "kube-api-access-vwl7v") pod "db33216f-b3e7-418f-ac05-129fb34f4538" (UID: "db33216f-b3e7-418f-ac05-129fb34f4538"). InnerVolumeSpecName "kube-api-access-vwl7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.658928 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config" (OuterVolumeSpecName: "config") pod "db33216f-b3e7-418f-ac05-129fb34f4538" (UID: "db33216f-b3e7-418f-ac05-129fb34f4538"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.658969 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "db33216f-b3e7-418f-ac05-129fb34f4538" (UID: "db33216f-b3e7-418f-ac05-129fb34f4538"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.676818 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "db33216f-b3e7-418f-ac05-129fb34f4538" (UID: "db33216f-b3e7-418f-ac05-129fb34f4538"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.680186 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "db33216f-b3e7-418f-ac05-129fb34f4538" (UID: "db33216f-b3e7-418f-ac05-129fb34f4538"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.704831 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.704878 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwl7v\" (UniqueName: \"kubernetes.io/projected/db33216f-b3e7-418f-ac05-129fb34f4538-kube-api-access-vwl7v\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.704889 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.704898 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:52 crc kubenswrapper[5002]: I1014 08:08:52.704907 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/db33216f-b3e7-418f-ac05-129fb34f4538-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044227 5002 generic.go:334] "Generic (PLEG): container finished" podID="db33216f-b3e7-418f-ac05-129fb34f4538" containerID="3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7" exitCode=0 Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044274 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" event={"ID":"db33216f-b3e7-418f-ac05-129fb34f4538","Type":"ContainerDied","Data":"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7"} Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044333 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" event={"ID":"db33216f-b3e7-418f-ac05-129fb34f4538","Type":"ContainerDied","Data":"c66098917b50f823128ff878ae39db3f8bbb1b3002b8d155f6b8591b1252170b"} Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044351 5002 scope.go:117] "RemoveContainer" containerID="3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044403 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="cinder-scheduler" containerID="cri-o://86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4" gracePeriod=30 Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044296 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f7955b85-fm7pg" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.044624 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="probe" containerID="cri-o://3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b" gracePeriod=30 Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.065350 5002 scope.go:117] "RemoveContainer" containerID="b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.077555 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.083071 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85f7955b85-fm7pg"] Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.092536 5002 scope.go:117] "RemoveContainer" containerID="3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7" Oct 14 08:08:53 crc kubenswrapper[5002]: E1014 08:08:53.092959 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7\": container with ID starting with 3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7 not found: ID does not exist" containerID="3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.093003 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7"} err="failed to get container status \"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7\": rpc error: code = NotFound desc = could not find container \"3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7\": container with ID starting with 3e3a53a7d612488558c60733c674e63c2b4c89f830bd5cede2074c2bf4f291d7 not found: ID does not exist" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.093029 5002 scope.go:117] "RemoveContainer" containerID="b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624" Oct 14 08:08:53 crc kubenswrapper[5002]: E1014 08:08:53.093301 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624\": container with ID starting with b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624 not found: ID does not exist" containerID="b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.093329 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624"} err="failed to get container status \"b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624\": rpc error: code = NotFound desc = could not find container \"b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624\": container with ID starting with b611265f9ef768442302e777e7afb6ece1d4e1d3b80fe16e94b9e420728ed624 not found: ID does not exist" Oct 14 08:08:53 crc kubenswrapper[5002]: I1014 08:08:53.730040 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" path="/var/lib/kubelet/pods/db33216f-b3e7-418f-ac05-129fb34f4538/volumes" Oct 14 08:08:54 crc kubenswrapper[5002]: I1014 08:08:54.056989 5002 generic.go:334] "Generic (PLEG): container finished" podID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerID="3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b" exitCode=0 Oct 14 08:08:54 crc kubenswrapper[5002]: I1014 08:08:54.057045 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerDied","Data":"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b"} Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.078700 5002 generic.go:334] "Generic (PLEG): container finished" podID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerID="9e71e0cf8fbe795bdbcf8d98fcb033b339f5649ea1fe370d253c12cd71a8129b" exitCode=0 Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.078758 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerDied","Data":"9e71e0cf8fbe795bdbcf8d98fcb033b339f5649ea1fe370d253c12cd71a8129b"} Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.222907 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.390445 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle\") pod \"d7091e2e-4bcf-48e7-814a-39179fad09d4\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.390520 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mglr\" (UniqueName: \"kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr\") pod \"d7091e2e-4bcf-48e7-814a-39179fad09d4\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.390601 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs\") pod \"d7091e2e-4bcf-48e7-814a-39179fad09d4\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.390699 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config\") pod \"d7091e2e-4bcf-48e7-814a-39179fad09d4\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.390807 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config\") pod \"d7091e2e-4bcf-48e7-814a-39179fad09d4\" (UID: \"d7091e2e-4bcf-48e7-814a-39179fad09d4\") " Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.406515 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr" (OuterVolumeSpecName: "kube-api-access-7mglr") pod "d7091e2e-4bcf-48e7-814a-39179fad09d4" (UID: "d7091e2e-4bcf-48e7-814a-39179fad09d4"). InnerVolumeSpecName "kube-api-access-7mglr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.419575 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d7091e2e-4bcf-48e7-814a-39179fad09d4" (UID: "d7091e2e-4bcf-48e7-814a-39179fad09d4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.444397 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config" (OuterVolumeSpecName: "config") pod "d7091e2e-4bcf-48e7-814a-39179fad09d4" (UID: "d7091e2e-4bcf-48e7-814a-39179fad09d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.459639 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7091e2e-4bcf-48e7-814a-39179fad09d4" (UID: "d7091e2e-4bcf-48e7-814a-39179fad09d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.484640 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d7091e2e-4bcf-48e7-814a-39179fad09d4" (UID: "d7091e2e-4bcf-48e7-814a-39179fad09d4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.492788 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.493077 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mglr\" (UniqueName: \"kubernetes.io/projected/d7091e2e-4bcf-48e7-814a-39179fad09d4-kube-api-access-7mglr\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.493231 5002 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.493349 5002 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.493509 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7091e2e-4bcf-48e7-814a-39179fad09d4-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:55 crc kubenswrapper[5002]: I1014 08:08:55.818786 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.001630 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.001762 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.001879 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.002021 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.002131 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.002196 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.002265 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhcl6\" (UniqueName: \"kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6\") pod \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\" (UID: \"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc\") " Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.003385 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.007505 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts" (OuterVolumeSpecName: "scripts") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.007549 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6" (OuterVolumeSpecName: "kube-api-access-qhcl6") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "kube-api-access-qhcl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.008937 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.059015 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.090791 5002 generic.go:334] "Generic (PLEG): container finished" podID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerID="86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4" exitCode=0 Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.090897 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.090907 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerDied","Data":"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4"} Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.091085 5002 scope.go:117] "RemoveContainer" containerID="3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.094539 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc","Type":"ContainerDied","Data":"d82c5946f200b3a7a7e0a5b41b5b0f6ae8192b984959e127f6eeda9c4216c9fa"} Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.101007 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cfc656d-2j56s" event={"ID":"d7091e2e-4bcf-48e7-814a-39179fad09d4","Type":"ContainerDied","Data":"119f1afe824d262b3b54e7c5c1608242b77e9ee11457880bed8ac6439eaf1ba0"} Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.101344 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cfc656d-2j56s" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.105122 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.105152 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.105165 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.105192 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhcl6\" (UniqueName: \"kubernetes.io/projected/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-kube-api-access-qhcl6\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.126980 5002 scope.go:117] "RemoveContainer" containerID="86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.133028 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.139728 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7cfc656d-2j56s"] Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.164129 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data" (OuterVolumeSpecName: "config-data") pod "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" (UID: "d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.165174 5002 scope.go:117] "RemoveContainer" containerID="3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.165619 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b\": container with ID starting with 3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b not found: ID does not exist" containerID="3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.165727 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b"} err="failed to get container status \"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b\": rpc error: code = NotFound desc = could not find container \"3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b\": container with ID starting with 3b13ecf034ef34e9e4100506ce94d12ddc815dd6183d4110c5603df91c4b5d8b not found: ID does not exist" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.165863 5002 scope.go:117] "RemoveContainer" containerID="86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.166238 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4\": container with ID starting with 86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4 not found: ID does not exist" containerID="86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.166341 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4"} err="failed to get container status \"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4\": rpc error: code = NotFound desc = could not find container \"86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4\": container with ID starting with 86539b05658b6c37ec410b824ae431f83cf02457991191990e411640fab9d2b4 not found: ID does not exist" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.166438 5002 scope.go:117] "RemoveContainer" containerID="69ae80016f8c5320fc50b1b9bce8c35caa9dcd2528dcd0d705d082389b37e845" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.192963 5002 scope.go:117] "RemoveContainer" containerID="9e71e0cf8fbe795bdbcf8d98fcb033b339f5649ea1fe370d253c12cd71a8129b" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.206677 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.427359 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.457944 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.467446 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468187 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="dnsmasq-dns" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468234 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="dnsmasq-dns" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468278 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="init" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468299 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="init" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468334 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-api" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468351 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-api" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468381 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45065e0e-2e90-427d-a49c-423e2ba340d7" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468397 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45065e0e-2e90-427d-a49c-423e2ba340d7" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468422 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468438 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468465 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="cinder-scheduler" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468481 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="cinder-scheduler" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468511 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a2af3f-d7e5-4e1e-89ba-b567ab162985" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468528 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a2af3f-d7e5-4e1e-89ba-b567ab162985" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468566 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-httpd" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468583 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-httpd" Oct 14 08:08:56 crc kubenswrapper[5002]: E1014 08:08:56.468622 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="probe" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.468639 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="probe" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469123 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="db33216f-b3e7-418f-ac05-129fb34f4538" containerName="dnsmasq-dns" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469206 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="45065e0e-2e90-427d-a49c-423e2ba340d7" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469238 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="probe" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469260 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469278 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-api" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469297 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" containerName="neutron-httpd" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469331 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a2af3f-d7e5-4e1e-89ba-b567ab162985" containerName="mariadb-database-create" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.469358 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" containerName="cinder-scheduler" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.471331 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.473321 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.477499 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613170 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3f1dbab-8a06-4b99-8f41-39179983cd6c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613576 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613691 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613734 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613841 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp54p\" (UniqueName: \"kubernetes.io/projected/f3f1dbab-8a06-4b99-8f41-39179983cd6c-kube-api-access-dp54p\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.613977 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.716329 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3f1dbab-8a06-4b99-8f41-39179983cd6c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.716399 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.716513 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.718007 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.716927 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f3f1dbab-8a06-4b99-8f41-39179983cd6c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.718141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp54p\" (UniqueName: \"kubernetes.io/projected/f3f1dbab-8a06-4b99-8f41-39179983cd6c-kube-api-access-dp54p\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.718264 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.722044 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.722456 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.723433 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-config-data\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.724007 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3f1dbab-8a06-4b99-8f41-39179983cd6c-scripts\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.737056 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp54p\" (UniqueName: \"kubernetes.io/projected/f3f1dbab-8a06-4b99-8f41-39179983cd6c-kube-api-access-dp54p\") pod \"cinder-scheduler-0\" (UID: \"f3f1dbab-8a06-4b99-8f41-39179983cd6c\") " pod="openstack/cinder-scheduler-0" Oct 14 08:08:56 crc kubenswrapper[5002]: I1014 08:08:56.802129 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.269728 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.650753 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9546-account-create-r4d9k"] Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.652244 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.655581 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.660594 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9546-account-create-r4d9k"] Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.730930 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7091e2e-4bcf-48e7-814a-39179fad09d4" path="/var/lib/kubelet/pods/d7091e2e-4bcf-48e7-814a-39179fad09d4/volumes" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.732163 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc" path="/var/lib/kubelet/pods/d9b01fe5-09f3-4644-94f5-7e5cd28a8ccc/volumes" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.834199 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-886c-account-create-sd5qd"] Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.835647 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.839243 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.842942 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-886c-account-create-sd5qd"] Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.846784 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chq7\" (UniqueName: \"kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7\") pod \"nova-api-9546-account-create-r4d9k\" (UID: \"d3ca6482-5693-4c7d-b49a-81b265362d6b\") " pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.948756 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvsff\" (UniqueName: \"kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff\") pod \"nova-cell0-886c-account-create-sd5qd\" (UID: \"e1037688-f513-41da-9b9b-ae374f5f7732\") " pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.948827 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chq7\" (UniqueName: \"kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7\") pod \"nova-api-9546-account-create-r4d9k\" (UID: \"d3ca6482-5693-4c7d-b49a-81b265362d6b\") " pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.966473 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chq7\" (UniqueName: \"kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7\") pod \"nova-api-9546-account-create-r4d9k\" (UID: \"d3ca6482-5693-4c7d-b49a-81b265362d6b\") " pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:08:57 crc kubenswrapper[5002]: I1014 08:08:57.980233 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.037357 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2a87-account-create-tmzd4"] Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.041573 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.044873 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.050513 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvsff\" (UniqueName: \"kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff\") pod \"nova-cell0-886c-account-create-sd5qd\" (UID: \"e1037688-f513-41da-9b9b-ae374f5f7732\") " pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.052359 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2a87-account-create-tmzd4"] Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.076145 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvsff\" (UniqueName: \"kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff\") pod \"nova-cell0-886c-account-create-sd5qd\" (UID: \"e1037688-f513-41da-9b9b-ae374f5f7732\") " pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.148337 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3f1dbab-8a06-4b99-8f41-39179983cd6c","Type":"ContainerStarted","Data":"a2eb65f9e83339339cb5b144abb57b4345758cf1ea3af84ecb77f1de699989c2"} Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.148396 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3f1dbab-8a06-4b99-8f41-39179983cd6c","Type":"ContainerStarted","Data":"6819cc5b8635ef7dc349f1496b6bd721936617e4c84e2cd46384e39df643be69"} Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.152073 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d544v\" (UniqueName: \"kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v\") pod \"nova-cell1-2a87-account-create-tmzd4\" (UID: \"f066883e-6310-43d3-8805-493589140410\") " pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.165498 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.194430 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.255624 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d544v\" (UniqueName: \"kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v\") pod \"nova-cell1-2a87-account-create-tmzd4\" (UID: \"f066883e-6310-43d3-8805-493589140410\") " pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.276498 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d544v\" (UniqueName: \"kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v\") pod \"nova-cell1-2a87-account-create-tmzd4\" (UID: \"f066883e-6310-43d3-8805-493589140410\") " pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.385509 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.495374 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9546-account-create-r4d9k"] Oct 14 08:08:58 crc kubenswrapper[5002]: W1014 08:08:58.516050 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3ca6482_5693_4c7d_b49a_81b265362d6b.slice/crio-6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02 WatchSource:0}: Error finding container 6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02: Status 404 returned error can't find the container with id 6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02 Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.650534 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-886c-account-create-sd5qd"] Oct 14 08:08:58 crc kubenswrapper[5002]: W1014 08:08:58.655553 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1037688_f513_41da_9b9b_ae374f5f7732.slice/crio-d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3 WatchSource:0}: Error finding container d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3: Status 404 returned error can't find the container with id d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3 Oct 14 08:08:58 crc kubenswrapper[5002]: I1014 08:08:58.879038 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2a87-account-create-tmzd4"] Oct 14 08:08:58 crc kubenswrapper[5002]: W1014 08:08:58.882313 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf066883e_6310_43d3_8805_493589140410.slice/crio-d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f WatchSource:0}: Error finding container d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f: Status 404 returned error can't find the container with id d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.161891 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f3f1dbab-8a06-4b99-8f41-39179983cd6c","Type":"ContainerStarted","Data":"56c80a013ae67c7b1f2d53ec8b5a67c8080918a8a60b43354ea4bf608c98e356"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.164428 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2a87-account-create-tmzd4" event={"ID":"f066883e-6310-43d3-8805-493589140410","Type":"ContainerStarted","Data":"17431fac5f13d50730a283bf8140e1a7cb93673360b6c9856973c3bfa780fec0"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.164460 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2a87-account-create-tmzd4" event={"ID":"f066883e-6310-43d3-8805-493589140410","Type":"ContainerStarted","Data":"d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.167997 5002 generic.go:334] "Generic (PLEG): container finished" podID="e1037688-f513-41da-9b9b-ae374f5f7732" containerID="563a3132c15960a9c0a81ec98db5e8ec19d59aa4551ce458c7f1194127a8629c" exitCode=0 Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.168166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-886c-account-create-sd5qd" event={"ID":"e1037688-f513-41da-9b9b-ae374f5f7732","Type":"ContainerDied","Data":"563a3132c15960a9c0a81ec98db5e8ec19d59aa4551ce458c7f1194127a8629c"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.168264 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-886c-account-create-sd5qd" event={"ID":"e1037688-f513-41da-9b9b-ae374f5f7732","Type":"ContainerStarted","Data":"d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.171067 5002 generic.go:334] "Generic (PLEG): container finished" podID="d3ca6482-5693-4c7d-b49a-81b265362d6b" containerID="d15712662a4a64b0e26d02543cec76a83451768b5bccfc80ab2ccf47811a35d2" exitCode=0 Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.171251 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9546-account-create-r4d9k" event={"ID":"d3ca6482-5693-4c7d-b49a-81b265362d6b","Type":"ContainerDied","Data":"d15712662a4a64b0e26d02543cec76a83451768b5bccfc80ab2ccf47811a35d2"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.171357 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9546-account-create-r4d9k" event={"ID":"d3ca6482-5693-4c7d-b49a-81b265362d6b","Type":"ContainerStarted","Data":"6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02"} Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.188718 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.188699104 podStartE2EDuration="3.188699104s" podCreationTimestamp="2025-10-14 08:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:59.183604238 +0000 UTC m=+1072.164843700" watchObservedRunningTime="2025-10-14 08:08:59.188699104 +0000 UTC m=+1072.169938566" Oct 14 08:08:59 crc kubenswrapper[5002]: I1014 08:08:59.244067 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-2a87-account-create-tmzd4" podStartSLOduration=1.244046429 podStartE2EDuration="1.244046429s" podCreationTimestamp="2025-10-14 08:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:08:59.23740521 +0000 UTC m=+1072.218644732" watchObservedRunningTime="2025-10-14 08:08:59.244046429 +0000 UTC m=+1072.225285891" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.184551 5002 generic.go:334] "Generic (PLEG): container finished" podID="f066883e-6310-43d3-8805-493589140410" containerID="17431fac5f13d50730a283bf8140e1a7cb93673360b6c9856973c3bfa780fec0" exitCode=0 Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.185117 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2a87-account-create-tmzd4" event={"ID":"f066883e-6310-43d3-8805-493589140410","Type":"ContainerDied","Data":"17431fac5f13d50730a283bf8140e1a7cb93673360b6c9856973c3bfa780fec0"} Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.639998 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.644883 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.840989 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvsff\" (UniqueName: \"kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff\") pod \"e1037688-f513-41da-9b9b-ae374f5f7732\" (UID: \"e1037688-f513-41da-9b9b-ae374f5f7732\") " Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.841157 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6chq7\" (UniqueName: \"kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7\") pod \"d3ca6482-5693-4c7d-b49a-81b265362d6b\" (UID: \"d3ca6482-5693-4c7d-b49a-81b265362d6b\") " Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.847662 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff" (OuterVolumeSpecName: "kube-api-access-lvsff") pod "e1037688-f513-41da-9b9b-ae374f5f7732" (UID: "e1037688-f513-41da-9b9b-ae374f5f7732"). InnerVolumeSpecName "kube-api-access-lvsff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.848362 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7" (OuterVolumeSpecName: "kube-api-access-6chq7") pod "d3ca6482-5693-4c7d-b49a-81b265362d6b" (UID: "d3ca6482-5693-4c7d-b49a-81b265362d6b"). InnerVolumeSpecName "kube-api-access-6chq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.943610 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6chq7\" (UniqueName: \"kubernetes.io/projected/d3ca6482-5693-4c7d-b49a-81b265362d6b-kube-api-access-6chq7\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:00 crc kubenswrapper[5002]: I1014 08:09:00.943666 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvsff\" (UniqueName: \"kubernetes.io/projected/e1037688-f513-41da-9b9b-ae374f5f7732-kube-api-access-lvsff\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.200169 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9546-account-create-r4d9k" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.200182 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9546-account-create-r4d9k" event={"ID":"d3ca6482-5693-4c7d-b49a-81b265362d6b","Type":"ContainerDied","Data":"6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02"} Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.200360 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fbe9d384d59616cbd8b6791e084f81667442b81a7b7ce84cfe9ac4c70fd1b02" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.204192 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-886c-account-create-sd5qd" event={"ID":"e1037688-f513-41da-9b9b-ae374f5f7732","Type":"ContainerDied","Data":"d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3"} Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.204239 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d07d707726e2c7e305ceb6a4a8697d66f2797d96fa7535fad9ae767212a13df3" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.204264 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-886c-account-create-sd5qd" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.595089 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.656935 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d544v\" (UniqueName: \"kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v\") pod \"f066883e-6310-43d3-8805-493589140410\" (UID: \"f066883e-6310-43d3-8805-493589140410\") " Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.663554 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v" (OuterVolumeSpecName: "kube-api-access-d544v") pod "f066883e-6310-43d3-8805-493589140410" (UID: "f066883e-6310-43d3-8805-493589140410"). InnerVolumeSpecName "kube-api-access-d544v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.759133 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d544v\" (UniqueName: \"kubernetes.io/projected/f066883e-6310-43d3-8805-493589140410-kube-api-access-d544v\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:01 crc kubenswrapper[5002]: I1014 08:09:01.802911 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 08:09:02 crc kubenswrapper[5002]: I1014 08:09:02.221229 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2a87-account-create-tmzd4" event={"ID":"f066883e-6310-43d3-8805-493589140410","Type":"ContainerDied","Data":"d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f"} Oct 14 08:09:02 crc kubenswrapper[5002]: I1014 08:09:02.222415 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d05a805939897e82a23bfd715486d64199f848f33c48b5bd88e4dbbaba61ea8f" Oct 14 08:09:02 crc kubenswrapper[5002]: I1014 08:09:02.221281 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2a87-account-create-tmzd4" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.064529 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jrn8p"] Oct 14 08:09:03 crc kubenswrapper[5002]: E1014 08:09:03.065228 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ca6482-5693-4c7d-b49a-81b265362d6b" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065240 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ca6482-5693-4c7d-b49a-81b265362d6b" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: E1014 08:09:03.065261 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f066883e-6310-43d3-8805-493589140410" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065267 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f066883e-6310-43d3-8805-493589140410" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: E1014 08:09:03.065277 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1037688-f513-41da-9b9b-ae374f5f7732" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065284 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1037688-f513-41da-9b9b-ae374f5f7732" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065445 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f066883e-6310-43d3-8805-493589140410" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065463 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1037688-f513-41da-9b9b-ae374f5f7732" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.065474 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ca6482-5693-4c7d-b49a-81b265362d6b" containerName="mariadb-account-create" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.066043 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.067923 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9kzpc" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.068149 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.068248 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.113128 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jrn8p"] Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.181890 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shr9h\" (UniqueName: \"kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.181948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.181982 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.182102 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.283770 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.283890 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.283973 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shr9h\" (UniqueName: \"kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.284053 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.289824 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.292013 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.310955 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.322600 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shr9h\" (UniqueName: \"kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h\") pod \"nova-cell0-conductor-db-sync-jrn8p\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.384064 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:03 crc kubenswrapper[5002]: I1014 08:09:03.886956 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jrn8p"] Oct 14 08:09:04 crc kubenswrapper[5002]: I1014 08:09:04.246517 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" event={"ID":"4ad94d62-d1f0-45b9-be90-7bf2b95612a9","Type":"ContainerStarted","Data":"853eada20d811f99efa9e7ca0e2f0b797e1290d633bbbe65403913085d5c63c7"} Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.532901 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.534628 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-central-agent" containerID="cri-o://01319d05524c3663aa130456f9f276c8bb2abc128cad32efd381546ea91b6e9f" gracePeriod=30 Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.534752 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="proxy-httpd" containerID="cri-o://a8079d269fb4a0b25a57bc735e82c3a922f2636e9704a9f1c002012c6594dc90" gracePeriod=30 Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.534791 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-notification-agent" containerID="cri-o://b636e5a146d90ec398735b2f28bffc2bdad71f267f9ba156161af9ee86493944" gracePeriod=30 Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.534768 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="sg-core" containerID="cri-o://6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892" gracePeriod=30 Oct 14 08:09:05 crc kubenswrapper[5002]: I1014 08:09:05.540158 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:09:05 crc kubenswrapper[5002]: E1014 08:09:05.643629 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod337d6a02_f0b2_4493_8950_0227aad2b92a.slice/crio-6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265222 5002 generic.go:334] "Generic (PLEG): container finished" podID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerID="a8079d269fb4a0b25a57bc735e82c3a922f2636e9704a9f1c002012c6594dc90" exitCode=0 Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265536 5002 generic.go:334] "Generic (PLEG): container finished" podID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerID="6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892" exitCode=2 Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265548 5002 generic.go:334] "Generic (PLEG): container finished" podID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerID="01319d05524c3663aa130456f9f276c8bb2abc128cad32efd381546ea91b6e9f" exitCode=0 Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265290 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerDied","Data":"a8079d269fb4a0b25a57bc735e82c3a922f2636e9704a9f1c002012c6594dc90"} Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265585 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerDied","Data":"6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892"} Oct 14 08:09:06 crc kubenswrapper[5002]: I1014 08:09:06.265600 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerDied","Data":"01319d05524c3663aa130456f9f276c8bb2abc128cad32efd381546ea91b6e9f"} Oct 14 08:09:07 crc kubenswrapper[5002]: I1014 08:09:07.192966 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 08:09:07 crc kubenswrapper[5002]: I1014 08:09:07.302074 5002 generic.go:334] "Generic (PLEG): container finished" podID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerID="b636e5a146d90ec398735b2f28bffc2bdad71f267f9ba156161af9ee86493944" exitCode=0 Oct 14 08:09:07 crc kubenswrapper[5002]: I1014 08:09:07.302120 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerDied","Data":"b636e5a146d90ec398735b2f28bffc2bdad71f267f9ba156161af9ee86493944"} Oct 14 08:09:08 crc kubenswrapper[5002]: I1014 08:09:08.448772 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:08 crc kubenswrapper[5002]: I1014 08:09:08.449253 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" containerName="kube-state-metrics" containerID="cri-o://459fafadcc02de6779287720ad9e5ceacb41d3138ae237077a4b199d56698627" gracePeriod=30 Oct 14 08:09:09 crc kubenswrapper[5002]: I1014 08:09:09.218145 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:09:09 crc kubenswrapper[5002]: I1014 08:09:09.218385 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:09:09 crc kubenswrapper[5002]: I1014 08:09:09.324392 5002 generic.go:334] "Generic (PLEG): container finished" podID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" containerID="459fafadcc02de6779287720ad9e5ceacb41d3138ae237077a4b199d56698627" exitCode=2 Oct 14 08:09:09 crc kubenswrapper[5002]: I1014 08:09:09.324457 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fd0a0da-37d6-424c-b8ad-87791d2b155e","Type":"ContainerDied","Data":"459fafadcc02de6779287720ad9e5ceacb41d3138ae237077a4b199d56698627"} Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.606154 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.669631 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.729191 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5vrq\" (UniqueName: \"kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq\") pod \"3fd0a0da-37d6-424c-b8ad-87791d2b155e\" (UID: \"3fd0a0da-37d6-424c-b8ad-87791d2b155e\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.738517 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq" (OuterVolumeSpecName: "kube-api-access-z5vrq") pod "3fd0a0da-37d6-424c-b8ad-87791d2b155e" (UID: "3fd0a0da-37d6-424c-b8ad-87791d2b155e"). InnerVolumeSpecName "kube-api-access-z5vrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.830508 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqs9l\" (UniqueName: \"kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.830596 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.830624 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.830758 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.830783 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.831045 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.831186 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd\") pod \"337d6a02-f0b2-4493-8950-0227aad2b92a\" (UID: \"337d6a02-f0b2-4493-8950-0227aad2b92a\") " Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.831260 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.831565 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.832091 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5vrq\" (UniqueName: \"kubernetes.io/projected/3fd0a0da-37d6-424c-b8ad-87791d2b155e-kube-api-access-z5vrq\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.832111 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.832120 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/337d6a02-f0b2-4493-8950-0227aad2b92a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.836126 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts" (OuterVolumeSpecName: "scripts") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.836274 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l" (OuterVolumeSpecName: "kube-api-access-xqs9l") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "kube-api-access-xqs9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.860082 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.898391 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.920496 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data" (OuterVolumeSpecName: "config-data") pod "337d6a02-f0b2-4493-8950-0227aad2b92a" (UID: "337d6a02-f0b2-4493-8950-0227aad2b92a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.934103 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqs9l\" (UniqueName: \"kubernetes.io/projected/337d6a02-f0b2-4493-8950-0227aad2b92a-kube-api-access-xqs9l\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.934142 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.934154 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.934166 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:11 crc kubenswrapper[5002]: I1014 08:09:11.934177 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337d6a02-f0b2-4493-8950-0227aad2b92a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.383255 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"337d6a02-f0b2-4493-8950-0227aad2b92a","Type":"ContainerDied","Data":"061023af8273d2e046acb34e23d26987c102a9695d9ef8a1da26e63b05620a83"} Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.383324 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.383337 5002 scope.go:117] "RemoveContainer" containerID="a8079d269fb4a0b25a57bc735e82c3a922f2636e9704a9f1c002012c6594dc90" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.386887 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fd0a0da-37d6-424c-b8ad-87791d2b155e","Type":"ContainerDied","Data":"c41ec304711e83e77f3aeab30b2209511ac236c34d4d76d7a17b44fba0a60ede"} Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.387118 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.392286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" event={"ID":"4ad94d62-d1f0-45b9-be90-7bf2b95612a9","Type":"ContainerStarted","Data":"ac189386219a24197085d751764c9963973dc00fbc6a7be042cda9abc8a28dce"} Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.410584 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" podStartSLOduration=1.9119185829999998 podStartE2EDuration="9.410566478s" podCreationTimestamp="2025-10-14 08:09:03 +0000 UTC" firstStartedPulling="2025-10-14 08:09:03.896956134 +0000 UTC m=+1076.878195586" lastFinishedPulling="2025-10-14 08:09:11.395604039 +0000 UTC m=+1084.376843481" observedRunningTime="2025-10-14 08:09:12.410056645 +0000 UTC m=+1085.391296097" watchObservedRunningTime="2025-10-14 08:09:12.410566478 +0000 UTC m=+1085.391805930" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.416551 5002 scope.go:117] "RemoveContainer" containerID="6d3fd52d669ce1276b762f059acf2cabc8f9eec68628753c75655d49fd825892" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.428139 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.432433 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.440105 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.446076 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455107 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: E1014 08:09:12.455451 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-notification-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455463 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-notification-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: E1014 08:09:12.455477 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="proxy-httpd" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455484 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="proxy-httpd" Oct 14 08:09:12 crc kubenswrapper[5002]: E1014 08:09:12.455495 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-central-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455501 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-central-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: E1014 08:09:12.455512 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" containerName="kube-state-metrics" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455519 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" containerName="kube-state-metrics" Oct 14 08:09:12 crc kubenswrapper[5002]: E1014 08:09:12.455541 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="sg-core" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455548 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="sg-core" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455705 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="sg-core" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455719 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" containerName="kube-state-metrics" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455730 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="proxy-httpd" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455738 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-notification-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.455745 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" containerName="ceilometer-central-agent" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.457236 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.459677 5002 scope.go:117] "RemoveContainer" containerID="b636e5a146d90ec398735b2f28bffc2bdad71f267f9ba156161af9ee86493944" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.461611 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.461792 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.461868 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vq5jr" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.462065 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.485378 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.486480 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.489730 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.489876 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.504048 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.514494 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.528934 5002 scope.go:117] "RemoveContainer" containerID="01319d05524c3663aa130456f9f276c8bb2abc128cad32efd381546ea91b6e9f" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.555863 5002 scope.go:117] "RemoveContainer" containerID="459fafadcc02de6779287720ad9e5ceacb41d3138ae237077a4b199d56698627" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646289 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh58p\" (UniqueName: \"kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646363 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646394 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l57h6\" (UniqueName: \"kubernetes.io/projected/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-api-access-l57h6\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646422 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646486 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646717 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646772 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.646950 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.647091 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.647114 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.647228 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749002 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749048 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749360 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749508 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749600 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749726 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749747 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh58p\" (UniqueName: \"kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749772 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l57h6\" (UniqueName: \"kubernetes.io/projected/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-api-access-l57h6\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749810 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749826 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.749852 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.750186 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.754096 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.755633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.755777 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.756162 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.756406 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.759199 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.760226 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.768653 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l57h6\" (UniqueName: \"kubernetes.io/projected/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-kube-api-access-l57h6\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.772156 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/670693df-bbd2-4bf4-a705-8cc8a02d3d2f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"670693df-bbd2-4bf4-a705-8cc8a02d3d2f\") " pod="openstack/kube-state-metrics-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.774025 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh58p\" (UniqueName: \"kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.775228 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.785526 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:12 crc kubenswrapper[5002]: I1014 08:09:12.826391 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.334767 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:13 crc kubenswrapper[5002]: W1014 08:09:13.339784 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61980a87_ebe6_4245_b464_4556f7c184dc.slice/crio-6cf479938b636be60ec7574b10dc8d0c2caa1e0f5022a5f5b8429f244de2ee77 WatchSource:0}: Error finding container 6cf479938b636be60ec7574b10dc8d0c2caa1e0f5022a5f5b8429f244de2ee77: Status 404 returned error can't find the container with id 6cf479938b636be60ec7574b10dc8d0c2caa1e0f5022a5f5b8429f244de2ee77 Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.368525 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 08:09:13 crc kubenswrapper[5002]: W1014 08:09:13.372865 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod670693df_bbd2_4bf4_a705_8cc8a02d3d2f.slice/crio-9db037ad6d107bf88489310091b1c829bdebb99eccb67d509702e3b2bb1a3489 WatchSource:0}: Error finding container 9db037ad6d107bf88489310091b1c829bdebb99eccb67d509702e3b2bb1a3489: Status 404 returned error can't find the container with id 9db037ad6d107bf88489310091b1c829bdebb99eccb67d509702e3b2bb1a3489 Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.425369 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerStarted","Data":"6cf479938b636be60ec7574b10dc8d0c2caa1e0f5022a5f5b8429f244de2ee77"} Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.427441 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"670693df-bbd2-4bf4-a705-8cc8a02d3d2f","Type":"ContainerStarted","Data":"9db037ad6d107bf88489310091b1c829bdebb99eccb67d509702e3b2bb1a3489"} Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.732655 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="337d6a02-f0b2-4493-8950-0227aad2b92a" path="/var/lib/kubelet/pods/337d6a02-f0b2-4493-8950-0227aad2b92a/volumes" Oct 14 08:09:13 crc kubenswrapper[5002]: I1014 08:09:13.733745 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fd0a0da-37d6-424c-b8ad-87791d2b155e" path="/var/lib/kubelet/pods/3fd0a0da-37d6-424c-b8ad-87791d2b155e/volumes" Oct 14 08:09:14 crc kubenswrapper[5002]: I1014 08:09:14.436458 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"670693df-bbd2-4bf4-a705-8cc8a02d3d2f","Type":"ContainerStarted","Data":"20c788b4f556e1f57e8771b393cfcee0ea063dd2fa38c4dc2f6883a774b85946"} Oct 14 08:09:14 crc kubenswrapper[5002]: I1014 08:09:14.436808 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 08:09:14 crc kubenswrapper[5002]: I1014 08:09:14.438229 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerStarted","Data":"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d"} Oct 14 08:09:14 crc kubenswrapper[5002]: I1014 08:09:14.458758 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.989278616 podStartE2EDuration="2.458740537s" podCreationTimestamp="2025-10-14 08:09:12 +0000 UTC" firstStartedPulling="2025-10-14 08:09:13.375147597 +0000 UTC m=+1086.356387049" lastFinishedPulling="2025-10-14 08:09:13.844609518 +0000 UTC m=+1086.825848970" observedRunningTime="2025-10-14 08:09:14.454968417 +0000 UTC m=+1087.436207889" watchObservedRunningTime="2025-10-14 08:09:14.458740537 +0000 UTC m=+1087.439979989" Oct 14 08:09:15 crc kubenswrapper[5002]: I1014 08:09:15.463121 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerStarted","Data":"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549"} Oct 14 08:09:16 crc kubenswrapper[5002]: I1014 08:09:16.473031 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerStarted","Data":"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705"} Oct 14 08:09:16 crc kubenswrapper[5002]: I1014 08:09:16.488117 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:09:16 crc kubenswrapper[5002]: I1014 08:09:16.728438 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5dc947dd58-kwj5n" Oct 14 08:09:17 crc kubenswrapper[5002]: I1014 08:09:17.514963 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerStarted","Data":"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223"} Oct 14 08:09:17 crc kubenswrapper[5002]: I1014 08:09:17.516554 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:09:17 crc kubenswrapper[5002]: I1014 08:09:17.554420 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.027458331 podStartE2EDuration="5.554401709s" podCreationTimestamp="2025-10-14 08:09:12 +0000 UTC" firstStartedPulling="2025-10-14 08:09:13.342122552 +0000 UTC m=+1086.323362004" lastFinishedPulling="2025-10-14 08:09:16.86906593 +0000 UTC m=+1089.850305382" observedRunningTime="2025-10-14 08:09:17.549489298 +0000 UTC m=+1090.530728770" watchObservedRunningTime="2025-10-14 08:09:17.554401709 +0000 UTC m=+1090.535641151" Oct 14 08:09:22 crc kubenswrapper[5002]: I1014 08:09:22.569100 5002 generic.go:334] "Generic (PLEG): container finished" podID="4ad94d62-d1f0-45b9-be90-7bf2b95612a9" containerID="ac189386219a24197085d751764c9963973dc00fbc6a7be042cda9abc8a28dce" exitCode=0 Oct 14 08:09:22 crc kubenswrapper[5002]: I1014 08:09:22.569885 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" event={"ID":"4ad94d62-d1f0-45b9-be90-7bf2b95612a9","Type":"ContainerDied","Data":"ac189386219a24197085d751764c9963973dc00fbc6a7be042cda9abc8a28dce"} Oct 14 08:09:22 crc kubenswrapper[5002]: I1014 08:09:22.839777 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 08:09:23 crc kubenswrapper[5002]: I1014 08:09:23.942798 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:23 crc kubenswrapper[5002]: I1014 08:09:23.996454 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle\") pod \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " Oct 14 08:09:23 crc kubenswrapper[5002]: I1014 08:09:23.996563 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data\") pod \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " Oct 14 08:09:23 crc kubenswrapper[5002]: I1014 08:09:23.996639 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shr9h\" (UniqueName: \"kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h\") pod \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " Oct 14 08:09:23 crc kubenswrapper[5002]: I1014 08:09:23.996672 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts\") pod \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\" (UID: \"4ad94d62-d1f0-45b9-be90-7bf2b95612a9\") " Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.002459 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts" (OuterVolumeSpecName: "scripts") pod "4ad94d62-d1f0-45b9-be90-7bf2b95612a9" (UID: "4ad94d62-d1f0-45b9-be90-7bf2b95612a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.002997 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h" (OuterVolumeSpecName: "kube-api-access-shr9h") pod "4ad94d62-d1f0-45b9-be90-7bf2b95612a9" (UID: "4ad94d62-d1f0-45b9-be90-7bf2b95612a9"). InnerVolumeSpecName "kube-api-access-shr9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.021394 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data" (OuterVolumeSpecName: "config-data") pod "4ad94d62-d1f0-45b9-be90-7bf2b95612a9" (UID: "4ad94d62-d1f0-45b9-be90-7bf2b95612a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.023664 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ad94d62-d1f0-45b9-be90-7bf2b95612a9" (UID: "4ad94d62-d1f0-45b9-be90-7bf2b95612a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.098761 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.098819 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.098891 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shr9h\" (UniqueName: \"kubernetes.io/projected/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-kube-api-access-shr9h\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.098916 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad94d62-d1f0-45b9-be90-7bf2b95612a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.592624 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" event={"ID":"4ad94d62-d1f0-45b9-be90-7bf2b95612a9","Type":"ContainerDied","Data":"853eada20d811f99efa9e7ca0e2f0b797e1290d633bbbe65403913085d5c63c7"} Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.592668 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jrn8p" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.592677 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="853eada20d811f99efa9e7ca0e2f0b797e1290d633bbbe65403913085d5c63c7" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.699168 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:24 crc kubenswrapper[5002]: E1014 08:09:24.699504 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad94d62-d1f0-45b9-be90-7bf2b95612a9" containerName="nova-cell0-conductor-db-sync" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.699520 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad94d62-d1f0-45b9-be90-7bf2b95612a9" containerName="nova-cell0-conductor-db-sync" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.699680 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad94d62-d1f0-45b9-be90-7bf2b95612a9" containerName="nova-cell0-conductor-db-sync" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.700225 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.705607 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9kzpc" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.708829 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.709582 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqgcz\" (UniqueName: \"kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.709800 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.709889 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.717450 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.812005 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqgcz\" (UniqueName: \"kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.812140 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.812177 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.820732 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.821080 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:24 crc kubenswrapper[5002]: I1014 08:09:24.827163 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqgcz\" (UniqueName: \"kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz\") pod \"nova-cell0-conductor-0\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:25 crc kubenswrapper[5002]: I1014 08:09:25.022933 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:25 crc kubenswrapper[5002]: I1014 08:09:25.487329 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:25 crc kubenswrapper[5002]: W1014 08:09:25.497878 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod956c650d_378c_4613_a3c8_ff022df8db07.slice/crio-6daa0be4ce67742b186b71b5596a8da19cbefd5d566a339eadb032b39da0ab5b WatchSource:0}: Error finding container 6daa0be4ce67742b186b71b5596a8da19cbefd5d566a339eadb032b39da0ab5b: Status 404 returned error can't find the container with id 6daa0be4ce67742b186b71b5596a8da19cbefd5d566a339eadb032b39da0ab5b Oct 14 08:09:25 crc kubenswrapper[5002]: I1014 08:09:25.607142 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"956c650d-378c-4613-a3c8-ff022df8db07","Type":"ContainerStarted","Data":"6daa0be4ce67742b186b71b5596a8da19cbefd5d566a339eadb032b39da0ab5b"} Oct 14 08:09:26 crc kubenswrapper[5002]: I1014 08:09:26.619252 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"956c650d-378c-4613-a3c8-ff022df8db07","Type":"ContainerStarted","Data":"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c"} Oct 14 08:09:26 crc kubenswrapper[5002]: I1014 08:09:26.619660 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:26 crc kubenswrapper[5002]: I1014 08:09:26.658027 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.6579967570000003 podStartE2EDuration="2.657996757s" podCreationTimestamp="2025-10-14 08:09:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:26.637628121 +0000 UTC m=+1099.618867643" watchObservedRunningTime="2025-10-14 08:09:26.657996757 +0000 UTC m=+1099.639236249" Oct 14 08:09:28 crc kubenswrapper[5002]: I1014 08:09:28.837566 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:28 crc kubenswrapper[5002]: I1014 08:09:28.839069 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="956c650d-378c-4613-a3c8-ff022df8db07" containerName="nova-cell0-conductor-conductor" containerID="cri-o://63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" gracePeriod=30 Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.028385 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c is running failed: container process not found" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.031114 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c is running failed: container process not found" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.031871 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c is running failed: container process not found" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.031930 5002 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="956c650d-378c-4613-a3c8-ff022df8db07" containerName="nova-cell0-conductor-conductor" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.035132 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.035357 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-central-agent" containerID="cri-o://0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d" gracePeriod=30 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.035981 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="proxy-httpd" containerID="cri-o://2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223" gracePeriod=30 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.036059 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="sg-core" containerID="cri-o://bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705" gracePeriod=30 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.036072 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-notification-agent" containerID="cri-o://58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549" gracePeriod=30 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.053504 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.171:3000/\": read tcp 10.217.0.2:34292->10.217.0.171:3000: read: connection reset by peer" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.113203 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.314252 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqgcz\" (UniqueName: \"kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz\") pod \"956c650d-378c-4613-a3c8-ff022df8db07\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.314737 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data\") pod \"956c650d-378c-4613-a3c8-ff022df8db07\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.314803 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle\") pod \"956c650d-378c-4613-a3c8-ff022df8db07\" (UID: \"956c650d-378c-4613-a3c8-ff022df8db07\") " Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.322243 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz" (OuterVolumeSpecName: "kube-api-access-mqgcz") pod "956c650d-378c-4613-a3c8-ff022df8db07" (UID: "956c650d-378c-4613-a3c8-ff022df8db07"). InnerVolumeSpecName "kube-api-access-mqgcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.354462 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "956c650d-378c-4613-a3c8-ff022df8db07" (UID: "956c650d-378c-4613-a3c8-ff022df8db07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.368485 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data" (OuterVolumeSpecName: "config-data") pod "956c650d-378c-4613-a3c8-ff022df8db07" (UID: "956c650d-378c-4613-a3c8-ff022df8db07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.416806 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqgcz\" (UniqueName: \"kubernetes.io/projected/956c650d-378c-4613-a3c8-ff022df8db07-kube-api-access-mqgcz\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.416876 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.416895 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956c650d-378c-4613-a3c8-ff022df8db07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.665883 5002 generic.go:334] "Generic (PLEG): container finished" podID="956c650d-378c-4613-a3c8-ff022df8db07" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" exitCode=0 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.665952 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"956c650d-378c-4613-a3c8-ff022df8db07","Type":"ContainerDied","Data":"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c"} Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.665982 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"956c650d-378c-4613-a3c8-ff022df8db07","Type":"ContainerDied","Data":"6daa0be4ce67742b186b71b5596a8da19cbefd5d566a339eadb032b39da0ab5b"} Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.666000 5002 scope.go:117] "RemoveContainer" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.666138 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673471 5002 generic.go:334] "Generic (PLEG): container finished" podID="61980a87-ebe6-4245-b464-4556f7c184dc" containerID="2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223" exitCode=0 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673495 5002 generic.go:334] "Generic (PLEG): container finished" podID="61980a87-ebe6-4245-b464-4556f7c184dc" containerID="bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705" exitCode=2 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673503 5002 generic.go:334] "Generic (PLEG): container finished" podID="61980a87-ebe6-4245-b464-4556f7c184dc" containerID="0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d" exitCode=0 Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673519 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerDied","Data":"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223"} Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673544 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerDied","Data":"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705"} Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.673554 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerDied","Data":"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d"} Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.713815 5002 scope.go:117] "RemoveContainer" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.714354 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c\": container with ID starting with 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c not found: ID does not exist" containerID="63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.714397 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.714404 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c"} err="failed to get container status \"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c\": rpc error: code = NotFound desc = could not find container \"63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c\": container with ID starting with 63cf28cf59d47d62020aa964c27de076212d1aeb854ae26764e770f6322f7c9c not found: ID does not exist" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.727758 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.748300 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:30 crc kubenswrapper[5002]: E1014 08:09:30.748697 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956c650d-378c-4613-a3c8-ff022df8db07" containerName="nova-cell0-conductor-conductor" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.748713 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="956c650d-378c-4613-a3c8-ff022df8db07" containerName="nova-cell0-conductor-conductor" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.748886 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="956c650d-378c-4613-a3c8-ff022df8db07" containerName="nova-cell0-conductor-conductor" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.749433 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.752986 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9kzpc" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.753026 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.757879 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.925675 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.926014 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f4zw\" (UniqueName: \"kubernetes.io/projected/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-kube-api-access-7f4zw\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:30 crc kubenswrapper[5002]: I1014 08:09:30.926242 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.028101 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f4zw\" (UniqueName: \"kubernetes.io/projected/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-kube-api-access-7f4zw\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.028293 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.028400 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.035374 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.035673 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.048058 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f4zw\" (UniqueName: \"kubernetes.io/projected/fc8e7450-29ac-4acb-a090-b7b0d1f2666f-kube-api-access-7f4zw\") pod \"nova-cell0-conductor-0\" (UID: \"fc8e7450-29ac-4acb-a090-b7b0d1f2666f\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.121030 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.136781 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333287 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333338 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333421 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh58p\" (UniqueName: \"kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333494 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333518 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333539 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333884 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.333944 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd\") pod \"61980a87-ebe6-4245-b464-4556f7c184dc\" (UID: \"61980a87-ebe6-4245-b464-4556f7c184dc\") " Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.334212 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.334758 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.334863 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.334879 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61980a87-ebe6-4245-b464-4556f7c184dc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.337484 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p" (OuterVolumeSpecName: "kube-api-access-lh58p") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "kube-api-access-lh58p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.338433 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts" (OuterVolumeSpecName: "scripts") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.372718 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.392654 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.406097 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.419137 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data" (OuterVolumeSpecName: "config-data") pod "61980a87-ebe6-4245-b464-4556f7c184dc" (UID: "61980a87-ebe6-4245-b464-4556f7c184dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436372 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh58p\" (UniqueName: \"kubernetes.io/projected/61980a87-ebe6-4245-b464-4556f7c184dc-kube-api-access-lh58p\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436412 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436428 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436441 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436453 5002 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.436465 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61980a87-ebe6-4245-b464-4556f7c184dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.580403 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.686310 5002 generic.go:334] "Generic (PLEG): container finished" podID="61980a87-ebe6-4245-b464-4556f7c184dc" containerID="58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549" exitCode=0 Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.686387 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerDied","Data":"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549"} Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.686419 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61980a87-ebe6-4245-b464-4556f7c184dc","Type":"ContainerDied","Data":"6cf479938b636be60ec7574b10dc8d0c2caa1e0f5022a5f5b8429f244de2ee77"} Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.686416 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.686459 5002 scope.go:117] "RemoveContainer" containerID="2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.688098 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fc8e7450-29ac-4acb-a090-b7b0d1f2666f","Type":"ContainerStarted","Data":"4d9051fbf5113e9635a6413fb32a0ad51253b2a919c7c24c46f0f46633df863c"} Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.710183 5002 scope.go:117] "RemoveContainer" containerID="bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.741150 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="956c650d-378c-4613-a3c8-ff022df8db07" path="/var/lib/kubelet/pods/956c650d-378c-4613-a3c8-ff022df8db07/volumes" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.742264 5002 scope.go:117] "RemoveContainer" containerID="58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.748564 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.759693 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.767805 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.768213 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-central-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768228 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-central-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.768244 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-notification-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768249 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-notification-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.768266 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="sg-core" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768273 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="sg-core" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.768286 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="proxy-httpd" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768292 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="proxy-httpd" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768435 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="sg-core" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768446 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="proxy-httpd" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768471 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-notification-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.768482 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" containerName="ceilometer-central-agent" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.769954 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.778201 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.778458 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.778613 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.780107 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.873558 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.873689 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.873733 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.873940 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.873977 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.874095 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.874128 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.874154 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqmw4\" (UniqueName: \"kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.877974 5002 scope.go:117] "RemoveContainer" containerID="0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.904147 5002 scope.go:117] "RemoveContainer" containerID="2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.906975 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223\": container with ID starting with 2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223 not found: ID does not exist" containerID="2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907004 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223"} err="failed to get container status \"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223\": rpc error: code = NotFound desc = could not find container \"2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223\": container with ID starting with 2fe70b2b821f146954f2a9a92621f85a06d4bbcdc00dbaa95bc059d867842223 not found: ID does not exist" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907023 5002 scope.go:117] "RemoveContainer" containerID="bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.907325 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705\": container with ID starting with bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705 not found: ID does not exist" containerID="bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907346 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705"} err="failed to get container status \"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705\": rpc error: code = NotFound desc = could not find container \"bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705\": container with ID starting with bd74ad0e992ab61f0ea124ab140b03c8ec67bd4591fc28a09a0f17c67b123705 not found: ID does not exist" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907359 5002 scope.go:117] "RemoveContainer" containerID="58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.907588 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549\": container with ID starting with 58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549 not found: ID does not exist" containerID="58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907610 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549"} err="failed to get container status \"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549\": rpc error: code = NotFound desc = could not find container \"58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549\": container with ID starting with 58abf6d4ebb83ffff0524d879c0712b4c37523723371710d3bd70ebe82eac549 not found: ID does not exist" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907622 5002 scope.go:117] "RemoveContainer" containerID="0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d" Oct 14 08:09:31 crc kubenswrapper[5002]: E1014 08:09:31.907829 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d\": container with ID starting with 0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d not found: ID does not exist" containerID="0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.907865 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d"} err="failed to get container status \"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d\": rpc error: code = NotFound desc = could not find container \"0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d\": container with ID starting with 0f73fc7b3096d35f29f8ba93b732ddcd66e2783b0d36b5006827ef939009b15d not found: ID does not exist" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976412 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976478 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976542 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976569 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976653 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976672 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqmw4\" (UniqueName: \"kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.976742 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.977875 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.978024 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.982144 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.983076 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.984017 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.984749 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.991482 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:31 crc kubenswrapper[5002]: I1014 08:09:31.994148 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqmw4\" (UniqueName: \"kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4\") pod \"ceilometer-0\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " pod="openstack/ceilometer-0" Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.183817 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.635741 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.708060 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fc8e7450-29ac-4acb-a090-b7b0d1f2666f","Type":"ContainerStarted","Data":"19e6d48c0b2c1bcf216c709c64a44a4e300197bfd0de1521835145b184063aac"} Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.709419 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.714581 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerStarted","Data":"63f36b6de4c610fd955703cbd349d03741f5759ad55b52b043f3cf55936572ed"} Oct 14 08:09:32 crc kubenswrapper[5002]: I1014 08:09:32.725769 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.725750889 podStartE2EDuration="2.725750889s" podCreationTimestamp="2025-10-14 08:09:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:32.724452113 +0000 UTC m=+1105.705691635" watchObservedRunningTime="2025-10-14 08:09:32.725750889 +0000 UTC m=+1105.706990351" Oct 14 08:09:33 crc kubenswrapper[5002]: I1014 08:09:33.743132 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61980a87-ebe6-4245-b464-4556f7c184dc" path="/var/lib/kubelet/pods/61980a87-ebe6-4245-b464-4556f7c184dc/volumes" Oct 14 08:09:33 crc kubenswrapper[5002]: I1014 08:09:33.746695 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerStarted","Data":"5d18d0e57fd977b82f23214d9fda6d136e76756a0c149dacc804f2f81af2952e"} Oct 14 08:09:34 crc kubenswrapper[5002]: I1014 08:09:34.756088 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerStarted","Data":"20c4859c26269158c99e23fa5b9c3ed79fd6b49fda3e81fb2b16b381d28aa7f3"} Oct 14 08:09:35 crc kubenswrapper[5002]: I1014 08:09:35.776350 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerStarted","Data":"64ddbdc17f84c52b6793ce6454c85d26ad30594b15e9be7d21e5c4239ca52252"} Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.179360 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.679424 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-kxtvt"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.697907 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.702028 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.702351 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.717621 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kxtvt"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.758243 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.758309 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.758338 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.758363 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xbjc\" (UniqueName: \"kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.797592 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerStarted","Data":"504bd83fd70888b4b45dee5862c8c70ec0a3e8c9ea3905bd60a47dcf35231316"} Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.798437 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.847474 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.848905 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.850668 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.859263 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.868095 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.869912 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870076 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870148 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xbjc\" (UniqueName: \"kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870354 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870471 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5shsr\" (UniqueName: \"kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870574 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.870737 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.871871 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.480275602 podStartE2EDuration="5.871852663s" podCreationTimestamp="2025-10-14 08:09:31 +0000 UTC" firstStartedPulling="2025-10-14 08:09:32.648007862 +0000 UTC m=+1105.629247314" lastFinishedPulling="2025-10-14 08:09:36.039584883 +0000 UTC m=+1109.020824375" observedRunningTime="2025-10-14 08:09:36.862484804 +0000 UTC m=+1109.843724256" watchObservedRunningTime="2025-10-14 08:09:36.871852663 +0000 UTC m=+1109.853092105" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.878094 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.879788 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.880590 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.917134 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.918527 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.919950 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xbjc\" (UniqueName: \"kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc\") pod \"nova-cell0-cell-mapping-kxtvt\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.924211 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.943623 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971234 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971275 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5shsr\" (UniqueName: \"kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971302 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971339 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971390 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971408 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.971485 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r9rn\" (UniqueName: \"kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.972006 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.978020 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.981945 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:36 crc kubenswrapper[5002]: I1014 08:09:36.994508 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5shsr\" (UniqueName: \"kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr\") pod \"nova-api-0\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " pod="openstack/nova-api-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.024418 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.027211 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.029087 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.032561 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.053822 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073763 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073822 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073879 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztfm5\" (UniqueName: \"kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073908 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073953 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.073992 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r9rn\" (UniqueName: \"kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.074026 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.078109 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.079207 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.081932 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.091359 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.093043 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.104389 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r9rn\" (UniqueName: \"kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn\") pod \"nova-cell1-novncproxy-0\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.117945 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177504 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jlkp\" (UniqueName: \"kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177586 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177654 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztfm5\" (UniqueName: \"kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177691 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177732 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177756 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.177806 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.178320 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.185611 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.189217 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.193528 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.201581 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.205851 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztfm5\" (UniqueName: \"kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5\") pod \"nova-metadata-0\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.207254 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.240579 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.292305 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305744 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305829 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305864 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76kwz\" (UniqueName: \"kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305890 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jlkp\" (UniqueName: \"kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305914 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305943 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.305976 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.306024 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.344460 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jlkp\" (UniqueName: \"kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.351400 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.356422 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.357592 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data\") pod \"nova-scheduler-0\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.409744 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.409809 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.409943 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.409961 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76kwz\" (UniqueName: \"kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.410001 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.410829 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.411889 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.412365 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.413496 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.446275 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76kwz\" (UniqueName: \"kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz\") pod \"dnsmasq-dns-5d8c9888b9-q7xlp\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.611621 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.692896 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.738493 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kxtvt"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.812877 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kxtvt" event={"ID":"4291140a-1f93-4552-bd10-ccbdb1b17b9f","Type":"ContainerStarted","Data":"2fbd6f87ddd7ad471696f8a748c39a71b5788a62e26d579ae585423996b54ede"} Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.860738 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: W1014 08:09:37.871134 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c20e8bb_24a3_477e_9a4a_414ac65c739a.slice/crio-c3af6dd41f726103a6b4d8a5f381ebbd1eb52ea07176bdac80bcc083e961515d WatchSource:0}: Error finding container c3af6dd41f726103a6b4d8a5f381ebbd1eb52ea07176bdac80bcc083e961515d: Status 404 returned error can't find the container with id c3af6dd41f726103a6b4d8a5f381ebbd1eb52ea07176bdac80bcc083e961515d Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.947484 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.958785 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.981721 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d5lkh"] Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.983501 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.989217 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:09:37 crc kubenswrapper[5002]: I1014 08:09:37.989407 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.005948 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d5lkh"] Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.028920 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72gtw\" (UniqueName: \"kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.028995 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.029042 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.029095 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.130506 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72gtw\" (UniqueName: \"kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.130564 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.130603 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.130641 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.135250 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.135566 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.136746 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.148874 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72gtw\" (UniqueName: \"kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw\") pod \"nova-cell1-conductor-db-sync-d5lkh\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.162573 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.312407 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.319276 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.827877 5002 generic.go:334] "Generic (PLEG): container finished" podID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerID="0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57" exitCode=0 Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.828535 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" event={"ID":"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4","Type":"ContainerDied","Data":"0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.828690 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" event={"ID":"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4","Type":"ContainerStarted","Data":"daff9bed9e51f47ee64fb18093936f5ae86b6f166598b88a5900da4a21a931d5"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.833674 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d00a3c1-1d35-4804-a2b1-268949d78c28","Type":"ContainerStarted","Data":"e783ecc94ebf0ab9452351f4bfcfe176510f2eb839ead8879b8aaf60d3acbb6c"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.846102 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerStarted","Data":"35a713eb2500cdc8434be9462414ac99935453903958452146f83f563aeb4eac"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.851036 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bcee502b-1550-42c9-8f5a-96482754b5e7","Type":"ContainerStarted","Data":"5b81050b4953202d58ac90e6fef84ddee8ef0fc840f55eae8611ccf4e9fb8e3b"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.852282 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerStarted","Data":"c3af6dd41f726103a6b4d8a5f381ebbd1eb52ea07176bdac80bcc083e961515d"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.857119 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d5lkh"] Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.864728 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kxtvt" event={"ID":"4291140a-1f93-4552-bd10-ccbdb1b17b9f","Type":"ContainerStarted","Data":"c0960d50784c055f2f1b3fd2f5be65005fdca52cd348b78dd0c07e41fff65106"} Oct 14 08:09:38 crc kubenswrapper[5002]: I1014 08:09:38.889195 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-kxtvt" podStartSLOduration=2.889175363 podStartE2EDuration="2.889175363s" podCreationTimestamp="2025-10-14 08:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:38.884185214 +0000 UTC m=+1111.865424666" watchObservedRunningTime="2025-10-14 08:09:38.889175363 +0000 UTC m=+1111.870414805" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.217888 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.218401 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.218473 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.219281 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.219348 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b" gracePeriod=600 Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.879572 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" event={"ID":"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4","Type":"ContainerStarted","Data":"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8"} Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.879960 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.884547 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b" exitCode=0 Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.884631 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b"} Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.884683 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5"} Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.884712 5002 scope.go:117] "RemoveContainer" containerID="599563e58039e5312be560a4192a79b8aa3ba55587514d7b8c7602992246525a" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.887071 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" event={"ID":"880bc121-fc8a-42e8-bb96-e1c0c615d39e","Type":"ContainerStarted","Data":"7bc84491255c430dc9c9c64476decd408acc72d7379ce1a3ed0fba09b11daf08"} Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.887138 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" event={"ID":"880bc121-fc8a-42e8-bb96-e1c0c615d39e","Type":"ContainerStarted","Data":"6e641b810299d49cc8aea877810ff1e50016f20615dd19c9b76ea5f5201f7fb4"} Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.896919 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" podStartSLOduration=2.896822694 podStartE2EDuration="2.896822694s" podCreationTimestamp="2025-10-14 08:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:39.89558425 +0000 UTC m=+1112.876823722" watchObservedRunningTime="2025-10-14 08:09:39.896822694 +0000 UTC m=+1112.878062146" Oct 14 08:09:39 crc kubenswrapper[5002]: I1014 08:09:39.933136 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" podStartSLOduration=2.933116865 podStartE2EDuration="2.933116865s" podCreationTimestamp="2025-10-14 08:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:39.92603834 +0000 UTC m=+1112.907277802" watchObservedRunningTime="2025-10-14 08:09:39.933116865 +0000 UTC m=+1112.914356317" Oct 14 08:09:40 crc kubenswrapper[5002]: I1014 08:09:40.888049 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:09:40 crc kubenswrapper[5002]: I1014 08:09:40.900797 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.925439 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerStarted","Data":"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe"} Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.932150 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d00a3c1-1d35-4804-a2b1-268949d78c28","Type":"ContainerStarted","Data":"7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d"} Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.953381 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bcee502b-1550-42c9-8f5a-96482754b5e7","Type":"ContainerStarted","Data":"1f01369e35849a39a781ed5d97212578aa7f7446321007e455568f44d3056513"} Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.953512 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bcee502b-1550-42c9-8f5a-96482754b5e7" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1f01369e35849a39a781ed5d97212578aa7f7446321007e455568f44d3056513" gracePeriod=30 Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.965134 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.5803399790000001 podStartE2EDuration="4.96510815s" podCreationTimestamp="2025-10-14 08:09:37 +0000 UTC" firstStartedPulling="2025-10-14 08:09:38.163987987 +0000 UTC m=+1111.145227439" lastFinishedPulling="2025-10-14 08:09:41.548756138 +0000 UTC m=+1114.529995610" observedRunningTime="2025-10-14 08:09:41.949868239 +0000 UTC m=+1114.931107691" watchObservedRunningTime="2025-10-14 08:09:41.96510815 +0000 UTC m=+1114.946347642" Oct 14 08:09:41 crc kubenswrapper[5002]: I1014 08:09:41.987314 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.388449314 podStartE2EDuration="5.987292663s" podCreationTimestamp="2025-10-14 08:09:36 +0000 UTC" firstStartedPulling="2025-10-14 08:09:37.956734097 +0000 UTC m=+1110.937973549" lastFinishedPulling="2025-10-14 08:09:41.555577446 +0000 UTC m=+1114.536816898" observedRunningTime="2025-10-14 08:09:41.967824625 +0000 UTC m=+1114.949064087" watchObservedRunningTime="2025-10-14 08:09:41.987292663 +0000 UTC m=+1114.968532115" Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.296888 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.612778 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.968594 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerStarted","Data":"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f"} Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.973070 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerStarted","Data":"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37"} Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.971901 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-metadata" containerID="cri-o://6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" gracePeriod=30 Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.971414 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-log" containerID="cri-o://b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" gracePeriod=30 Oct 14 08:09:42 crc kubenswrapper[5002]: I1014 08:09:42.983590 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerStarted","Data":"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d"} Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.034825 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.362316282 podStartE2EDuration="7.034800804s" podCreationTimestamp="2025-10-14 08:09:36 +0000 UTC" firstStartedPulling="2025-10-14 08:09:37.874097906 +0000 UTC m=+1110.855337358" lastFinishedPulling="2025-10-14 08:09:41.546582418 +0000 UTC m=+1114.527821880" observedRunningTime="2025-10-14 08:09:43.031178884 +0000 UTC m=+1116.012418346" watchObservedRunningTime="2025-10-14 08:09:43.034800804 +0000 UTC m=+1116.016040286" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.041031 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.47640036 podStartE2EDuration="7.041009765s" podCreationTimestamp="2025-10-14 08:09:36 +0000 UTC" firstStartedPulling="2025-10-14 08:09:37.983181087 +0000 UTC m=+1110.964420539" lastFinishedPulling="2025-10-14 08:09:41.547790472 +0000 UTC m=+1114.529029944" observedRunningTime="2025-10-14 08:09:43.014031391 +0000 UTC m=+1115.995270863" watchObservedRunningTime="2025-10-14 08:09:43.041009765 +0000 UTC m=+1116.022249247" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.585658 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.766004 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztfm5\" (UniqueName: \"kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5\") pod \"249267b9-bc98-42d6-b339-390bf1ad5bd7\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.766229 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data\") pod \"249267b9-bc98-42d6-b339-390bf1ad5bd7\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.766266 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle\") pod \"249267b9-bc98-42d6-b339-390bf1ad5bd7\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.766323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs\") pod \"249267b9-bc98-42d6-b339-390bf1ad5bd7\" (UID: \"249267b9-bc98-42d6-b339-390bf1ad5bd7\") " Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.766969 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs" (OuterVolumeSpecName: "logs") pod "249267b9-bc98-42d6-b339-390bf1ad5bd7" (UID: "249267b9-bc98-42d6-b339-390bf1ad5bd7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.767347 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/249267b9-bc98-42d6-b339-390bf1ad5bd7-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.787577 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5" (OuterVolumeSpecName: "kube-api-access-ztfm5") pod "249267b9-bc98-42d6-b339-390bf1ad5bd7" (UID: "249267b9-bc98-42d6-b339-390bf1ad5bd7"). InnerVolumeSpecName "kube-api-access-ztfm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.811327 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "249267b9-bc98-42d6-b339-390bf1ad5bd7" (UID: "249267b9-bc98-42d6-b339-390bf1ad5bd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.814810 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data" (OuterVolumeSpecName: "config-data") pod "249267b9-bc98-42d6-b339-390bf1ad5bd7" (UID: "249267b9-bc98-42d6-b339-390bf1ad5bd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.868932 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.868965 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249267b9-bc98-42d6-b339-390bf1ad5bd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.870466 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztfm5\" (UniqueName: \"kubernetes.io/projected/249267b9-bc98-42d6-b339-390bf1ad5bd7-kube-api-access-ztfm5\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.995868 5002 generic.go:334] "Generic (PLEG): container finished" podID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerID="6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" exitCode=0 Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.996243 5002 generic.go:334] "Generic (PLEG): container finished" podID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerID="b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" exitCode=143 Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.997634 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.998017 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerDied","Data":"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37"} Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.998072 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerDied","Data":"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f"} Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.998087 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"249267b9-bc98-42d6-b339-390bf1ad5bd7","Type":"ContainerDied","Data":"35a713eb2500cdc8434be9462414ac99935453903958452146f83f563aeb4eac"} Oct 14 08:09:43 crc kubenswrapper[5002]: I1014 08:09:43.998107 5002 scope.go:117] "RemoveContainer" containerID="6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.020353 5002 scope.go:117] "RemoveContainer" containerID="b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.061566 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.070675 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.076483 5002 scope.go:117] "RemoveContainer" containerID="6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" Oct 14 08:09:44 crc kubenswrapper[5002]: E1014 08:09:44.077691 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37\": container with ID starting with 6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37 not found: ID does not exist" containerID="6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.077718 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37"} err="failed to get container status \"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37\": rpc error: code = NotFound desc = could not find container \"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37\": container with ID starting with 6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37 not found: ID does not exist" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.077738 5002 scope.go:117] "RemoveContainer" containerID="b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" Oct 14 08:09:44 crc kubenswrapper[5002]: E1014 08:09:44.079480 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f\": container with ID starting with b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f not found: ID does not exist" containerID="b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.079548 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f"} err="failed to get container status \"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f\": rpc error: code = NotFound desc = could not find container \"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f\": container with ID starting with b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f not found: ID does not exist" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.079588 5002 scope.go:117] "RemoveContainer" containerID="6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.080085 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37"} err="failed to get container status \"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37\": rpc error: code = NotFound desc = could not find container \"6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37\": container with ID starting with 6d28d8282cbc8f616b6d4f9eff56f008312e78751cf27ebca8404d375859cc37 not found: ID does not exist" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.080153 5002 scope.go:117] "RemoveContainer" containerID="b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.080344 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:44 crc kubenswrapper[5002]: E1014 08:09:44.081694 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-log" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.081718 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-log" Oct 14 08:09:44 crc kubenswrapper[5002]: E1014 08:09:44.081741 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-metadata" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.081748 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-metadata" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.082204 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-log" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.082242 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" containerName="nova-metadata-metadata" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.082728 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f"} err="failed to get container status \"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f\": rpc error: code = NotFound desc = could not find container \"b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f\": container with ID starting with b80acb908f6de4472758088d42b67c230fe96b0b57cb6ca1250bfac31ef3609f not found: ID does not exist" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.083461 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.095599 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.095849 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.117761 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.178096 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.178172 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.178198 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.178219 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vdvt\" (UniqueName: \"kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.178247 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.279088 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.279297 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.279378 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vdvt\" (UniqueName: \"kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.279565 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.279673 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.280278 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.285789 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.286117 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.289231 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.306299 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vdvt\" (UniqueName: \"kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt\") pod \"nova-metadata-0\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.418814 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:44 crc kubenswrapper[5002]: I1014 08:09:44.928133 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:44 crc kubenswrapper[5002]: W1014 08:09:44.955100 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1479b1a2_0882_4e7c_9939_2500d14aeacf.slice/crio-0849023d90dd38330c71cd626962895985f2ca833bac4bede873afbae7eb0773 WatchSource:0}: Error finding container 0849023d90dd38330c71cd626962895985f2ca833bac4bede873afbae7eb0773: Status 404 returned error can't find the container with id 0849023d90dd38330c71cd626962895985f2ca833bac4bede873afbae7eb0773 Oct 14 08:09:45 crc kubenswrapper[5002]: I1014 08:09:45.005070 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerStarted","Data":"0849023d90dd38330c71cd626962895985f2ca833bac4bede873afbae7eb0773"} Oct 14 08:09:45 crc kubenswrapper[5002]: I1014 08:09:45.731941 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="249267b9-bc98-42d6-b339-390bf1ad5bd7" path="/var/lib/kubelet/pods/249267b9-bc98-42d6-b339-390bf1ad5bd7/volumes" Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.019490 5002 generic.go:334] "Generic (PLEG): container finished" podID="880bc121-fc8a-42e8-bb96-e1c0c615d39e" containerID="7bc84491255c430dc9c9c64476decd408acc72d7379ce1a3ed0fba09b11daf08" exitCode=0 Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.019658 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" event={"ID":"880bc121-fc8a-42e8-bb96-e1c0c615d39e","Type":"ContainerDied","Data":"7bc84491255c430dc9c9c64476decd408acc72d7379ce1a3ed0fba09b11daf08"} Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.024678 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerStarted","Data":"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6"} Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.024738 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerStarted","Data":"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6"} Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.026691 5002 generic.go:334] "Generic (PLEG): container finished" podID="4291140a-1f93-4552-bd10-ccbdb1b17b9f" containerID="c0960d50784c055f2f1b3fd2f5be65005fdca52cd348b78dd0c07e41fff65106" exitCode=0 Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.026815 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kxtvt" event={"ID":"4291140a-1f93-4552-bd10-ccbdb1b17b9f","Type":"ContainerDied","Data":"c0960d50784c055f2f1b3fd2f5be65005fdca52cd348b78dd0c07e41fff65106"} Oct 14 08:09:46 crc kubenswrapper[5002]: I1014 08:09:46.084954 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.08493541 podStartE2EDuration="2.08493541s" podCreationTimestamp="2025-10-14 08:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:46.06830686 +0000 UTC m=+1119.049546332" watchObservedRunningTime="2025-10-14 08:09:46.08493541 +0000 UTC m=+1119.066174862" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.187509 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.187900 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.589553 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.595740 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.612740 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.649887 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72gtw\" (UniqueName: \"kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw\") pod \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.649952 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle\") pod \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650050 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data\") pod \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650084 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data\") pod \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650186 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle\") pod \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650263 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts\") pod \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\" (UID: \"880bc121-fc8a-42e8-bb96-e1c0c615d39e\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650287 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xbjc\" (UniqueName: \"kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc\") pod \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.650402 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts\") pod \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\" (UID: \"4291140a-1f93-4552-bd10-ccbdb1b17b9f\") " Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.654244 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.656900 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts" (OuterVolumeSpecName: "scripts") pod "880bc121-fc8a-42e8-bb96-e1c0c615d39e" (UID: "880bc121-fc8a-42e8-bb96-e1c0c615d39e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.657035 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw" (OuterVolumeSpecName: "kube-api-access-72gtw") pod "880bc121-fc8a-42e8-bb96-e1c0c615d39e" (UID: "880bc121-fc8a-42e8-bb96-e1c0c615d39e"). InnerVolumeSpecName "kube-api-access-72gtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.664137 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc" (OuterVolumeSpecName: "kube-api-access-5xbjc") pod "4291140a-1f93-4552-bd10-ccbdb1b17b9f" (UID: "4291140a-1f93-4552-bd10-ccbdb1b17b9f"). InnerVolumeSpecName "kube-api-access-5xbjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.666729 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts" (OuterVolumeSpecName: "scripts") pod "4291140a-1f93-4552-bd10-ccbdb1b17b9f" (UID: "4291140a-1f93-4552-bd10-ccbdb1b17b9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.690621 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data" (OuterVolumeSpecName: "config-data") pod "880bc121-fc8a-42e8-bb96-e1c0c615d39e" (UID: "880bc121-fc8a-42e8-bb96-e1c0c615d39e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.694919 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.696071 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "880bc121-fc8a-42e8-bb96-e1c0c615d39e" (UID: "880bc121-fc8a-42e8-bb96-e1c0c615d39e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.698125 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data" (OuterVolumeSpecName: "config-data") pod "4291140a-1f93-4552-bd10-ccbdb1b17b9f" (UID: "4291140a-1f93-4552-bd10-ccbdb1b17b9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.704687 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4291140a-1f93-4552-bd10-ccbdb1b17b9f" (UID: "4291140a-1f93-4552-bd10-ccbdb1b17b9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753684 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753724 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753736 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753746 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xbjc\" (UniqueName: \"kubernetes.io/projected/4291140a-1f93-4552-bd10-ccbdb1b17b9f-kube-api-access-5xbjc\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753755 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4291140a-1f93-4552-bd10-ccbdb1b17b9f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753764 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72gtw\" (UniqueName: \"kubernetes.io/projected/880bc121-fc8a-42e8-bb96-e1c0c615d39e-kube-api-access-72gtw\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753772 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.753780 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880bc121-fc8a-42e8-bb96-e1c0c615d39e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.772239 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:09:47 crc kubenswrapper[5002]: I1014 08:09:47.775431 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="dnsmasq-dns" containerID="cri-o://422d62c0af860c269d29d2f2a479c116db633b500492cd3c298a01ec8cdc3e89" gracePeriod=10 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.051690 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" event={"ID":"880bc121-fc8a-42e8-bb96-e1c0c615d39e","Type":"ContainerDied","Data":"6e641b810299d49cc8aea877810ff1e50016f20615dd19c9b76ea5f5201f7fb4"} Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.051965 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e641b810299d49cc8aea877810ff1e50016f20615dd19c9b76ea5f5201f7fb4" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.051965 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-d5lkh" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.061818 5002 generic.go:334] "Generic (PLEG): container finished" podID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerID="422d62c0af860c269d29d2f2a479c116db633b500492cd3c298a01ec8cdc3e89" exitCode=0 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.061892 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" event={"ID":"6946ff35-58ec-4de0-be57-f077bbb50c63","Type":"ContainerDied","Data":"422d62c0af860c269d29d2f2a479c116db633b500492cd3c298a01ec8cdc3e89"} Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.064486 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kxtvt" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.064510 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kxtvt" event={"ID":"4291140a-1f93-4552-bd10-ccbdb1b17b9f","Type":"ContainerDied","Data":"2fbd6f87ddd7ad471696f8a748c39a71b5788a62e26d579ae585423996b54ede"} Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.064541 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fbd6f87ddd7ad471696f8a748c39a71b5788a62e26d579ae585423996b54ede" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.095352 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.127182 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: E1014 08:09:48.127577 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4291140a-1f93-4552-bd10-ccbdb1b17b9f" containerName="nova-manage" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.127592 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4291140a-1f93-4552-bd10-ccbdb1b17b9f" containerName="nova-manage" Oct 14 08:09:48 crc kubenswrapper[5002]: E1014 08:09:48.127607 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="880bc121-fc8a-42e8-bb96-e1c0c615d39e" containerName="nova-cell1-conductor-db-sync" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.127615 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="880bc121-fc8a-42e8-bb96-e1c0c615d39e" containerName="nova-cell1-conductor-db-sync" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.127775 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4291140a-1f93-4552-bd10-ccbdb1b17b9f" containerName="nova-manage" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.127791 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="880bc121-fc8a-42e8-bb96-e1c0c615d39e" containerName="nova-cell1-conductor-db-sync" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.128381 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.131764 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.158530 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.162672 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.162887 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.162918 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwft7\" (UniqueName: \"kubernetes.io/projected/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-kube-api-access-hwft7\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.185632 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.266509 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb\") pod \"6946ff35-58ec-4de0-be57-f077bbb50c63\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.266592 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8c8p\" (UniqueName: \"kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p\") pod \"6946ff35-58ec-4de0-be57-f077bbb50c63\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.266742 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config\") pod \"6946ff35-58ec-4de0-be57-f077bbb50c63\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.266769 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc\") pod \"6946ff35-58ec-4de0-be57-f077bbb50c63\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.266817 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb\") pod \"6946ff35-58ec-4de0-be57-f077bbb50c63\" (UID: \"6946ff35-58ec-4de0-be57-f077bbb50c63\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.267069 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.267091 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwft7\" (UniqueName: \"kubernetes.io/projected/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-kube-api-access-hwft7\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.267140 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.270931 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.271429 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.281452 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.284280 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.284400 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.284468 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-log" containerID="cri-o://34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe" gracePeriod=30 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.284499 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-api" containerID="cri-o://a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d" gracePeriod=30 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.285013 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p" (OuterVolumeSpecName: "kube-api-access-v8c8p") pod "6946ff35-58ec-4de0-be57-f077bbb50c63" (UID: "6946ff35-58ec-4de0-be57-f077bbb50c63"). InnerVolumeSpecName "kube-api-access-v8c8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.313419 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.313616 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-log" containerID="cri-o://928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" gracePeriod=30 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.313753 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-metadata" containerID="cri-o://d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" gracePeriod=30 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.329324 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwft7\" (UniqueName: \"kubernetes.io/projected/ba1b8fb3-2b74-47f6-adfe-58a6084311fb-kube-api-access-hwft7\") pod \"nova-cell1-conductor-0\" (UID: \"ba1b8fb3-2b74-47f6-adfe-58a6084311fb\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.362581 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6946ff35-58ec-4de0-be57-f077bbb50c63" (UID: "6946ff35-58ec-4de0-be57-f077bbb50c63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.368608 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.368631 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8c8p\" (UniqueName: \"kubernetes.io/projected/6946ff35-58ec-4de0-be57-f077bbb50c63-kube-api-access-v8c8p\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.374229 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6946ff35-58ec-4de0-be57-f077bbb50c63" (UID: "6946ff35-58ec-4de0-be57-f077bbb50c63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.381874 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6946ff35-58ec-4de0-be57-f077bbb50c63" (UID: "6946ff35-58ec-4de0-be57-f077bbb50c63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.383426 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config" (OuterVolumeSpecName: "config") pod "6946ff35-58ec-4de0-be57-f077bbb50c63" (UID: "6946ff35-58ec-4de0-be57-f077bbb50c63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.454802 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.469762 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.470135 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.470146 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6946ff35-58ec-4de0-be57-f077bbb50c63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.581804 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: W1014 08:09:48.909376 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba1b8fb3_2b74_47f6_adfe_58a6084311fb.slice/crio-9ceffbb32dac971623527f66fe8c58de552887cf1703cd8b0f3a98a6473d4318 WatchSource:0}: Error finding container 9ceffbb32dac971623527f66fe8c58de552887cf1703cd8b0f3a98a6473d4318: Status 404 returned error can't find the container with id 9ceffbb32dac971623527f66fe8c58de552887cf1703cd8b0f3a98a6473d4318 Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.914897 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.952143 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.977050 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs\") pod \"1479b1a2-0882-4e7c-9939-2500d14aeacf\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.977095 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs\") pod \"1479b1a2-0882-4e7c-9939-2500d14aeacf\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.977127 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data\") pod \"1479b1a2-0882-4e7c-9939-2500d14aeacf\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.977144 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vdvt\" (UniqueName: \"kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt\") pod \"1479b1a2-0882-4e7c-9939-2500d14aeacf\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.977175 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle\") pod \"1479b1a2-0882-4e7c-9939-2500d14aeacf\" (UID: \"1479b1a2-0882-4e7c-9939-2500d14aeacf\") " Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.978205 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs" (OuterVolumeSpecName: "logs") pod "1479b1a2-0882-4e7c-9939-2500d14aeacf" (UID: "1479b1a2-0882-4e7c-9939-2500d14aeacf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:48 crc kubenswrapper[5002]: I1014 08:09:48.988041 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt" (OuterVolumeSpecName: "kube-api-access-4vdvt") pod "1479b1a2-0882-4e7c-9939-2500d14aeacf" (UID: "1479b1a2-0882-4e7c-9939-2500d14aeacf"). InnerVolumeSpecName "kube-api-access-4vdvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.038418 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data" (OuterVolumeSpecName: "config-data") pod "1479b1a2-0882-4e7c-9939-2500d14aeacf" (UID: "1479b1a2-0882-4e7c-9939-2500d14aeacf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.040158 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1479b1a2-0882-4e7c-9939-2500d14aeacf" (UID: "1479b1a2-0882-4e7c-9939-2500d14aeacf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.041730 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1479b1a2-0882-4e7c-9939-2500d14aeacf" (UID: "1479b1a2-0882-4e7c-9939-2500d14aeacf"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.084584 5002 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.084620 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1479b1a2-0882-4e7c-9939-2500d14aeacf-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.084641 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.084655 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vdvt\" (UniqueName: \"kubernetes.io/projected/1479b1a2-0882-4e7c-9939-2500d14aeacf-kube-api-access-4vdvt\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.084668 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1479b1a2-0882-4e7c-9939-2500d14aeacf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085436 5002 generic.go:334] "Generic (PLEG): container finished" podID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerID="d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" exitCode=0 Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085469 5002 generic.go:334] "Generic (PLEG): container finished" podID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerID="928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" exitCode=143 Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085534 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerDied","Data":"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085570 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerDied","Data":"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085586 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1479b1a2-0882-4e7c-9939-2500d14aeacf","Type":"ContainerDied","Data":"0849023d90dd38330c71cd626962895985f2ca833bac4bede873afbae7eb0773"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085619 5002 scope.go:117] "RemoveContainer" containerID="d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.085828 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.089178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" event={"ID":"6946ff35-58ec-4de0-be57-f077bbb50c63","Type":"ContainerDied","Data":"d8d43b40b822bb4c7e60152ae106293ecd216e826a7b5c1daf6601d9af7bb8d6"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.089243 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68fcd76675-6nqbq" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.097479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerDied","Data":"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.097453 5002 generic.go:334] "Generic (PLEG): container finished" podID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerID="34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe" exitCode=143 Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.102352 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ba1b8fb3-2b74-47f6-adfe-58a6084311fb","Type":"ContainerStarted","Data":"9ceffbb32dac971623527f66fe8c58de552887cf1703cd8b0f3a98a6473d4318"} Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.142485 5002 scope.go:117] "RemoveContainer" containerID="928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.142591 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.153224 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68fcd76675-6nqbq"] Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.161431 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.172506 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.182763 5002 scope.go:117] "RemoveContainer" containerID="d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184295 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.184699 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="dnsmasq-dns" Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.184703 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6\": container with ID starting with d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6 not found: ID does not exist" containerID="d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184743 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6"} err="failed to get container status \"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6\": rpc error: code = NotFound desc = could not find container \"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6\": container with ID starting with d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6 not found: ID does not exist" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184772 5002 scope.go:117] "RemoveContainer" containerID="928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184716 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="dnsmasq-dns" Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.184914 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-log" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184931 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-log" Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.184941 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="init" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.184950 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="init" Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.185010 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-metadata" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.185019 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-metadata" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.185714 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" containerName="dnsmasq-dns" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.185746 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-metadata" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.185762 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" containerName="nova-metadata-log" Oct 14 08:09:49 crc kubenswrapper[5002]: E1014 08:09:49.186070 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6\": container with ID starting with 928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6 not found: ID does not exist" containerID="928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.186141 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6"} err="failed to get container status \"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6\": rpc error: code = NotFound desc = could not find container \"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6\": container with ID starting with 928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6 not found: ID does not exist" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.186336 5002 scope.go:117] "RemoveContainer" containerID="d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.186693 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6"} err="failed to get container status \"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6\": rpc error: code = NotFound desc = could not find container \"d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6\": container with ID starting with d842b84909feb1cb93f11feeddae720d9e1985eb6bad8f368c1b2831f93f4af6 not found: ID does not exist" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.186724 5002 scope.go:117] "RemoveContainer" containerID="928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.186991 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6"} err="failed to get container status \"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6\": rpc error: code = NotFound desc = could not find container \"928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6\": container with ID starting with 928af749b02bef258f2a3ec6750eb46602f60cf0a2d35f003e20fbb881669ec6 not found: ID does not exist" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.187012 5002 scope.go:117] "RemoveContainer" containerID="422d62c0af860c269d29d2f2a479c116db633b500492cd3c298a01ec8cdc3e89" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.187743 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.189815 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.190005 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.194182 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.212606 5002 scope.go:117] "RemoveContainer" containerID="0de5d4de7980754fcab606ab40622637524494c1bbc322adeb7ade3b688bd60d" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.287143 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.287224 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.287334 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.287412 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.287433 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st257\" (UniqueName: \"kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.389279 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.389339 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.389364 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st257\" (UniqueName: \"kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.389432 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.389451 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.390735 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.395740 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.396358 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.396481 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.410283 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st257\" (UniqueName: \"kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257\") pod \"nova-metadata-0\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.512175 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.741951 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1479b1a2-0882-4e7c-9939-2500d14aeacf" path="/var/lib/kubelet/pods/1479b1a2-0882-4e7c-9939-2500d14aeacf/volumes" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.743256 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6946ff35-58ec-4de0-be57-f077bbb50c63" path="/var/lib/kubelet/pods/6946ff35-58ec-4de0-be57-f077bbb50c63/volumes" Oct 14 08:09:49 crc kubenswrapper[5002]: I1014 08:09:49.962874 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:09:49 crc kubenswrapper[5002]: W1014 08:09:49.978147 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26720557_b8fa_4460_bf5f_1922335bb86b.slice/crio-9d40e60617616e1150d8b8fb388e1836aa71ddcb2c8aed5b566e486cc5f5a668 WatchSource:0}: Error finding container 9d40e60617616e1150d8b8fb388e1836aa71ddcb2c8aed5b566e486cc5f5a668: Status 404 returned error can't find the container with id 9d40e60617616e1150d8b8fb388e1836aa71ddcb2c8aed5b566e486cc5f5a668 Oct 14 08:09:50 crc kubenswrapper[5002]: I1014 08:09:50.125147 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerStarted","Data":"9d40e60617616e1150d8b8fb388e1836aa71ddcb2c8aed5b566e486cc5f5a668"} Oct 14 08:09:50 crc kubenswrapper[5002]: I1014 08:09:50.127566 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerName="nova-scheduler-scheduler" containerID="cri-o://7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" gracePeriod=30 Oct 14 08:09:50 crc kubenswrapper[5002]: I1014 08:09:50.128471 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ba1b8fb3-2b74-47f6-adfe-58a6084311fb","Type":"ContainerStarted","Data":"0130e8da9578426519e19ceb9aac0486f5024accebb822687ddb337c0998a662"} Oct 14 08:09:50 crc kubenswrapper[5002]: I1014 08:09:50.128500 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:50 crc kubenswrapper[5002]: I1014 08:09:50.160342 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.160321333 podStartE2EDuration="2.160321333s" podCreationTimestamp="2025-10-14 08:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:50.15515684 +0000 UTC m=+1123.136396312" watchObservedRunningTime="2025-10-14 08:09:50.160321333 +0000 UTC m=+1123.141560795" Oct 14 08:09:51 crc kubenswrapper[5002]: I1014 08:09:51.136497 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerStarted","Data":"58e93834475a6a8014a5b86627c2f7fc7e4429dbe8060f057453badbb7ded9ba"} Oct 14 08:09:51 crc kubenswrapper[5002]: I1014 08:09:51.137266 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerStarted","Data":"6589b31335c69bd2566d93b8a7ec5e5cf35be148c7597bfca9a9c50992feda9e"} Oct 14 08:09:51 crc kubenswrapper[5002]: I1014 08:09:51.164871 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.164828668 podStartE2EDuration="2.164828668s" podCreationTimestamp="2025-10-14 08:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:51.158127273 +0000 UTC m=+1124.139366755" watchObservedRunningTime="2025-10-14 08:09:51.164828668 +0000 UTC m=+1124.146068130" Oct 14 08:09:52 crc kubenswrapper[5002]: E1014 08:09:52.613812 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:09:52 crc kubenswrapper[5002]: E1014 08:09:52.615449 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:09:52 crc kubenswrapper[5002]: E1014 08:09:52.616700 5002 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 08:09:52 crc kubenswrapper[5002]: E1014 08:09:52.616779 5002 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerName="nova-scheduler-scheduler" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.165395 5002 generic.go:334] "Generic (PLEG): container finished" podID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerID="7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" exitCode=0 Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.165541 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d00a3c1-1d35-4804-a2b1-268949d78c28","Type":"ContainerDied","Data":"7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d"} Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.525387 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.571059 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jlkp\" (UniqueName: \"kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp\") pod \"6d00a3c1-1d35-4804-a2b1-268949d78c28\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.572198 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle\") pod \"6d00a3c1-1d35-4804-a2b1-268949d78c28\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.572316 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data\") pod \"6d00a3c1-1d35-4804-a2b1-268949d78c28\" (UID: \"6d00a3c1-1d35-4804-a2b1-268949d78c28\") " Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.582013 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp" (OuterVolumeSpecName: "kube-api-access-6jlkp") pod "6d00a3c1-1d35-4804-a2b1-268949d78c28" (UID: "6d00a3c1-1d35-4804-a2b1-268949d78c28"). InnerVolumeSpecName "kube-api-access-6jlkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.596389 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data" (OuterVolumeSpecName: "config-data") pod "6d00a3c1-1d35-4804-a2b1-268949d78c28" (UID: "6d00a3c1-1d35-4804-a2b1-268949d78c28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.611676 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d00a3c1-1d35-4804-a2b1-268949d78c28" (UID: "6d00a3c1-1d35-4804-a2b1-268949d78c28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.674293 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jlkp\" (UniqueName: \"kubernetes.io/projected/6d00a3c1-1d35-4804-a2b1-268949d78c28-kube-api-access-6jlkp\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.674331 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.674340 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d00a3c1-1d35-4804-a2b1-268949d78c28-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:53 crc kubenswrapper[5002]: I1014 08:09:53.994005 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.090092 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs\") pod \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.090180 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle\") pod \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.090226 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data\") pod \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.090308 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5shsr\" (UniqueName: \"kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr\") pod \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\" (UID: \"7c20e8bb-24a3-477e-9a4a-414ac65c739a\") " Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.090721 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs" (OuterVolumeSpecName: "logs") pod "7c20e8bb-24a3-477e-9a4a-414ac65c739a" (UID: "7c20e8bb-24a3-477e-9a4a-414ac65c739a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.093830 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr" (OuterVolumeSpecName: "kube-api-access-5shsr") pod "7c20e8bb-24a3-477e-9a4a-414ac65c739a" (UID: "7c20e8bb-24a3-477e-9a4a-414ac65c739a"). InnerVolumeSpecName "kube-api-access-5shsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.110915 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c20e8bb-24a3-477e-9a4a-414ac65c739a" (UID: "7c20e8bb-24a3-477e-9a4a-414ac65c739a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.129062 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data" (OuterVolumeSpecName: "config-data") pod "7c20e8bb-24a3-477e-9a4a-414ac65c739a" (UID: "7c20e8bb-24a3-477e-9a4a-414ac65c739a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.175689 5002 generic.go:334] "Generic (PLEG): container finished" podID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerID="a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d" exitCode=0 Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.175765 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerDied","Data":"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d"} Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.175788 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.175894 5002 scope.go:117] "RemoveContainer" containerID="a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.175872 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7c20e8bb-24a3-477e-9a4a-414ac65c739a","Type":"ContainerDied","Data":"c3af6dd41f726103a6b4d8a5f381ebbd1eb52ea07176bdac80bcc083e961515d"} Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.183535 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.183541 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d00a3c1-1d35-4804-a2b1-268949d78c28","Type":"ContainerDied","Data":"e783ecc94ebf0ab9452351f4bfcfe176510f2eb839ead8879b8aaf60d3acbb6c"} Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.192044 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5shsr\" (UniqueName: \"kubernetes.io/projected/7c20e8bb-24a3-477e-9a4a-414ac65c739a-kube-api-access-5shsr\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.192071 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c20e8bb-24a3-477e-9a4a-414ac65c739a-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.192081 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.192090 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c20e8bb-24a3-477e-9a4a-414ac65c739a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.257055 5002 scope.go:117] "RemoveContainer" containerID="34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.283147 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.301614 5002 scope.go:117] "RemoveContainer" containerID="a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.301756 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: E1014 08:09:54.304095 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d\": container with ID starting with a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d not found: ID does not exist" containerID="a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.304146 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d"} err="failed to get container status \"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d\": rpc error: code = NotFound desc = could not find container \"a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d\": container with ID starting with a5cc19f903d4d2c61951d92dedcdd29a9108da08113011643b369fa51947ff3d not found: ID does not exist" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.304193 5002 scope.go:117] "RemoveContainer" containerID="34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe" Oct 14 08:09:54 crc kubenswrapper[5002]: E1014 08:09:54.307262 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe\": container with ID starting with 34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe not found: ID does not exist" containerID="34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.307308 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe"} err="failed to get container status \"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe\": rpc error: code = NotFound desc = could not find container \"34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe\": container with ID starting with 34fa54722301b177df0e361b53236a799f90dcc557e269824c5cdf4d605cedfe not found: ID does not exist" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.307337 5002 scope.go:117] "RemoveContainer" containerID="7fedebce78e15c0d690ce6701cad8ccef8e8bf6ef9fa8792d76acc66057f3c3d" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.323564 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.336802 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.351696 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: E1014 08:09:54.352427 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-log" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352449 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-log" Oct 14 08:09:54 crc kubenswrapper[5002]: E1014 08:09:54.352490 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerName="nova-scheduler-scheduler" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352498 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerName="nova-scheduler-scheduler" Oct 14 08:09:54 crc kubenswrapper[5002]: E1014 08:09:54.352534 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-api" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352542 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-api" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352783 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-api" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352826 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" containerName="nova-scheduler-scheduler" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.352861 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" containerName="nova-api-log" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.353637 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.355132 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.361954 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.371412 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.372914 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.378299 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.379472 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395382 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395438 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395471 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395674 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395913 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kbh8\" (UniqueName: \"kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.395954 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.396153 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfrdz\" (UniqueName: \"kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.497899 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498002 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498060 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498106 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498252 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kbh8\" (UniqueName: \"kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498273 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498373 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfrdz\" (UniqueName: \"kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.498545 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.502097 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.502532 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.503477 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.512485 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.512549 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.513291 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.520269 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfrdz\" (UniqueName: \"kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz\") pod \"nova-api-0\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " pod="openstack/nova-api-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.521571 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kbh8\" (UniqueName: \"kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8\") pod \"nova-scheduler-0\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.668821 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:09:54 crc kubenswrapper[5002]: I1014 08:09:54.689513 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:09:55 crc kubenswrapper[5002]: I1014 08:09:55.167847 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:09:55 crc kubenswrapper[5002]: I1014 08:09:55.195499 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6e0662d7-dc68-4c29-a61a-64b03c15057b","Type":"ContainerStarted","Data":"bcdaf945492057679227b1cc7a41e8cce3eb868d7725e33b5403a6f7a2eae40a"} Oct 14 08:09:55 crc kubenswrapper[5002]: I1014 08:09:55.241695 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:09:55 crc kubenswrapper[5002]: W1014 08:09:55.243520 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podabb4c773_a284_4c5e_8f93_4b023241adaf.slice/crio-6f024419f653b7ab40bc587bad7b8ef7aa41a25ed061d8b8603ff423f5e63607 WatchSource:0}: Error finding container 6f024419f653b7ab40bc587bad7b8ef7aa41a25ed061d8b8603ff423f5e63607: Status 404 returned error can't find the container with id 6f024419f653b7ab40bc587bad7b8ef7aa41a25ed061d8b8603ff423f5e63607 Oct 14 08:09:55 crc kubenswrapper[5002]: I1014 08:09:55.737764 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d00a3c1-1d35-4804-a2b1-268949d78c28" path="/var/lib/kubelet/pods/6d00a3c1-1d35-4804-a2b1-268949d78c28/volumes" Oct 14 08:09:55 crc kubenswrapper[5002]: I1014 08:09:55.738428 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c20e8bb-24a3-477e-9a4a-414ac65c739a" path="/var/lib/kubelet/pods/7c20e8bb-24a3-477e-9a4a-414ac65c739a/volumes" Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.212274 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6e0662d7-dc68-4c29-a61a-64b03c15057b","Type":"ContainerStarted","Data":"b846ebb53cb1875c506c96afc2d86693fc36142011ea7fd373ba8eef43afa869"} Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.215290 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerStarted","Data":"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958"} Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.215338 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerStarted","Data":"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d"} Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.215356 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerStarted","Data":"6f024419f653b7ab40bc587bad7b8ef7aa41a25ed061d8b8603ff423f5e63607"} Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.260020 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.259984227 podStartE2EDuration="2.259984227s" podCreationTimestamp="2025-10-14 08:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:56.246224977 +0000 UTC m=+1129.227464479" watchObservedRunningTime="2025-10-14 08:09:56.259984227 +0000 UTC m=+1129.241223749" Oct 14 08:09:56 crc kubenswrapper[5002]: I1014 08:09:56.273782 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.273762237 podStartE2EDuration="2.273762237s" podCreationTimestamp="2025-10-14 08:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:09:56.271067773 +0000 UTC m=+1129.252307295" watchObservedRunningTime="2025-10-14 08:09:56.273762237 +0000 UTC m=+1129.255001699" Oct 14 08:09:58 crc kubenswrapper[5002]: I1014 08:09:58.488212 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 08:09:59 crc kubenswrapper[5002]: I1014 08:09:59.513097 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:09:59 crc kubenswrapper[5002]: I1014 08:09:59.513190 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:09:59 crc kubenswrapper[5002]: I1014 08:09:59.669701 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:10:00 crc kubenswrapper[5002]: I1014 08:10:00.524155 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:00 crc kubenswrapper[5002]: I1014 08:10:00.524169 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:02 crc kubenswrapper[5002]: I1014 08:10:02.199916 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:10:04 crc kubenswrapper[5002]: I1014 08:10:04.670190 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 08:10:04 crc kubenswrapper[5002]: I1014 08:10:04.692166 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:10:04 crc kubenswrapper[5002]: I1014 08:10:04.692231 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:10:04 crc kubenswrapper[5002]: I1014 08:10:04.718817 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 08:10:05 crc kubenswrapper[5002]: I1014 08:10:05.374372 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 08:10:05 crc kubenswrapper[5002]: I1014 08:10:05.774123 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:05 crc kubenswrapper[5002]: I1014 08:10:05.774100 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:09 crc kubenswrapper[5002]: I1014 08:10:09.522036 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:10:09 crc kubenswrapper[5002]: I1014 08:10:09.523716 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:10:09 crc kubenswrapper[5002]: I1014 08:10:09.530612 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:10:10 crc kubenswrapper[5002]: I1014 08:10:10.401974 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.418703 5002 generic.go:334] "Generic (PLEG): container finished" podID="bcee502b-1550-42c9-8f5a-96482754b5e7" containerID="1f01369e35849a39a781ed5d97212578aa7f7446321007e455568f44d3056513" exitCode=137 Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.418823 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bcee502b-1550-42c9-8f5a-96482754b5e7","Type":"ContainerDied","Data":"1f01369e35849a39a781ed5d97212578aa7f7446321007e455568f44d3056513"} Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.419146 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bcee502b-1550-42c9-8f5a-96482754b5e7","Type":"ContainerDied","Data":"5b81050b4953202d58ac90e6fef84ddee8ef0fc840f55eae8611ccf4e9fb8e3b"} Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.419171 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b81050b4953202d58ac90e6fef84ddee8ef0fc840f55eae8611ccf4e9fb8e3b" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.490794 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.570795 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data\") pod \"bcee502b-1550-42c9-8f5a-96482754b5e7\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.570910 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r9rn\" (UniqueName: \"kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn\") pod \"bcee502b-1550-42c9-8f5a-96482754b5e7\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.571131 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle\") pod \"bcee502b-1550-42c9-8f5a-96482754b5e7\" (UID: \"bcee502b-1550-42c9-8f5a-96482754b5e7\") " Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.576444 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn" (OuterVolumeSpecName: "kube-api-access-5r9rn") pod "bcee502b-1550-42c9-8f5a-96482754b5e7" (UID: "bcee502b-1550-42c9-8f5a-96482754b5e7"). InnerVolumeSpecName "kube-api-access-5r9rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.596353 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcee502b-1550-42c9-8f5a-96482754b5e7" (UID: "bcee502b-1550-42c9-8f5a-96482754b5e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.608331 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data" (OuterVolumeSpecName: "config-data") pod "bcee502b-1550-42c9-8f5a-96482754b5e7" (UID: "bcee502b-1550-42c9-8f5a-96482754b5e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.674061 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.674090 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcee502b-1550-42c9-8f5a-96482754b5e7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:12 crc kubenswrapper[5002]: I1014 08:10:12.674101 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r9rn\" (UniqueName: \"kubernetes.io/projected/bcee502b-1550-42c9-8f5a-96482754b5e7-kube-api-access-5r9rn\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.430697 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.484618 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.497866 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.517870 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:10:13 crc kubenswrapper[5002]: E1014 08:10:13.518589 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee502b-1550-42c9-8f5a-96482754b5e7" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.518624 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee502b-1550-42c9-8f5a-96482754b5e7" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.519048 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcee502b-1550-42c9-8f5a-96482754b5e7" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.520337 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.524887 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.524981 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.525047 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.531621 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.587653 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.588003 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.588140 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df8k8\" (UniqueName: \"kubernetes.io/projected/e9b0062d-6204-46b8-8614-26f00ff2efd7-kube-api-access-df8k8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.588287 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.588421 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.690861 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.690966 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.691010 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.691032 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df8k8\" (UniqueName: \"kubernetes.io/projected/e9b0062d-6204-46b8-8614-26f00ff2efd7-kube-api-access-df8k8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.691081 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.697624 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.698289 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.699779 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.701611 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9b0062d-6204-46b8-8614-26f00ff2efd7-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.712364 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df8k8\" (UniqueName: \"kubernetes.io/projected/e9b0062d-6204-46b8-8614-26f00ff2efd7-kube-api-access-df8k8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9b0062d-6204-46b8-8614-26f00ff2efd7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.732885 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcee502b-1550-42c9-8f5a-96482754b5e7" path="/var/lib/kubelet/pods/bcee502b-1550-42c9-8f5a-96482754b5e7/volumes" Oct 14 08:10:13 crc kubenswrapper[5002]: I1014 08:10:13.855710 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.392531 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.442488 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9b0062d-6204-46b8-8614-26f00ff2efd7","Type":"ContainerStarted","Data":"ad57146569a92fbb920de2f470bf121383336cf24da905b149b38c56f5715792"} Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.697631 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.699381 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.700044 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:10:14 crc kubenswrapper[5002]: I1014 08:10:14.708260 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.459714 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9b0062d-6204-46b8-8614-26f00ff2efd7","Type":"ContainerStarted","Data":"134028b3127f7b66a87fba8de37e937d92b4e4e599d9424e1f09bae0c548e0ea"} Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.460193 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.478338 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.492212 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.492189307 podStartE2EDuration="2.492189307s" podCreationTimestamp="2025-10-14 08:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:15.485965896 +0000 UTC m=+1148.467205388" watchObservedRunningTime="2025-10-14 08:10:15.492189307 +0000 UTC m=+1148.473428789" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.733149 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.734772 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.744027 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4fkk\" (UniqueName: \"kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.744142 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.744194 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.744275 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.744302 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.745530 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.846471 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4fkk\" (UniqueName: \"kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.846567 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.846610 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.846678 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.846709 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.847827 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.848067 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.848181 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.848406 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:15 crc kubenswrapper[5002]: I1014 08:10:15.885580 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4fkk\" (UniqueName: \"kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk\") pod \"dnsmasq-dns-6f9cf55b5c-fpcpc\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:16 crc kubenswrapper[5002]: I1014 08:10:16.062922 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:16 crc kubenswrapper[5002]: I1014 08:10:16.525797 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.476610 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5891d56-c935-449d-a120-14ebe283db57" containerID="bb471d6a417baa5671dfb6daaf368b02bd70846e31713d8f0ce7cfd7697c3f55" exitCode=0 Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.476723 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" event={"ID":"c5891d56-c935-449d-a120-14ebe283db57","Type":"ContainerDied","Data":"bb471d6a417baa5671dfb6daaf368b02bd70846e31713d8f0ce7cfd7697c3f55"} Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.476980 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" event={"ID":"c5891d56-c935-449d-a120-14ebe283db57","Type":"ContainerStarted","Data":"fb1dbea78883a1644d9bc2b0ff01e3f54d1ded8178ec3c28794a65d5c6ea7b9c"} Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.883484 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.884081 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-central-agent" containerID="cri-o://5d18d0e57fd977b82f23214d9fda6d136e76756a0c149dacc804f2f81af2952e" gracePeriod=30 Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.884217 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="proxy-httpd" containerID="cri-o://504bd83fd70888b4b45dee5862c8c70ec0a3e8c9ea3905bd60a47dcf35231316" gracePeriod=30 Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.884264 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="sg-core" containerID="cri-o://64ddbdc17f84c52b6793ce6454c85d26ad30594b15e9be7d21e5c4239ca52252" gracePeriod=30 Oct 14 08:10:17 crc kubenswrapper[5002]: I1014 08:10:17.884309 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-notification-agent" containerID="cri-o://20c4859c26269158c99e23fa5b9c3ed79fd6b49fda3e81fb2b16b381d28aa7f3" gracePeriod=30 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.110007 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487648 5002 generic.go:334] "Generic (PLEG): container finished" podID="41221834-182a-406d-9b98-26a51af2f6d7" containerID="504bd83fd70888b4b45dee5862c8c70ec0a3e8c9ea3905bd60a47dcf35231316" exitCode=0 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487921 5002 generic.go:334] "Generic (PLEG): container finished" podID="41221834-182a-406d-9b98-26a51af2f6d7" containerID="64ddbdc17f84c52b6793ce6454c85d26ad30594b15e9be7d21e5c4239ca52252" exitCode=2 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487928 5002 generic.go:334] "Generic (PLEG): container finished" podID="41221834-182a-406d-9b98-26a51af2f6d7" containerID="5d18d0e57fd977b82f23214d9fda6d136e76756a0c149dacc804f2f81af2952e" exitCode=0 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487961 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerDied","Data":"504bd83fd70888b4b45dee5862c8c70ec0a3e8c9ea3905bd60a47dcf35231316"} Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487985 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerDied","Data":"64ddbdc17f84c52b6793ce6454c85d26ad30594b15e9be7d21e5c4239ca52252"} Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.487994 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerDied","Data":"5d18d0e57fd977b82f23214d9fda6d136e76756a0c149dacc804f2f81af2952e"} Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.489555 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-log" containerID="cri-o://f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d" gracePeriod=30 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.489998 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" event={"ID":"c5891d56-c935-449d-a120-14ebe283db57","Type":"ContainerStarted","Data":"63115e709943833ad05e7ec6407a64046c52be7d78a17239489c100bd50817af"} Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.490414 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-api" containerID="cri-o://79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958" gracePeriod=30 Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.490599 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.510295 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" podStartSLOduration=3.510279219 podStartE2EDuration="3.510279219s" podCreationTimestamp="2025-10-14 08:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:18.507502352 +0000 UTC m=+1151.488741824" watchObservedRunningTime="2025-10-14 08:10:18.510279219 +0000 UTC m=+1151.491518671" Oct 14 08:10:18 crc kubenswrapper[5002]: I1014 08:10:18.856107 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:19 crc kubenswrapper[5002]: I1014 08:10:19.500804 5002 generic.go:334] "Generic (PLEG): container finished" podID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerID="f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d" exitCode=143 Oct 14 08:10:19 crc kubenswrapper[5002]: I1014 08:10:19.500872 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerDied","Data":"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d"} Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.181672 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.273635 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfrdz\" (UniqueName: \"kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz\") pod \"abb4c773-a284-4c5e-8f93-4b023241adaf\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.273764 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs\") pod \"abb4c773-a284-4c5e-8f93-4b023241adaf\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.273892 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle\") pod \"abb4c773-a284-4c5e-8f93-4b023241adaf\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.273917 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data\") pod \"abb4c773-a284-4c5e-8f93-4b023241adaf\" (UID: \"abb4c773-a284-4c5e-8f93-4b023241adaf\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.276027 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs" (OuterVolumeSpecName: "logs") pod "abb4c773-a284-4c5e-8f93-4b023241adaf" (UID: "abb4c773-a284-4c5e-8f93-4b023241adaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.282052 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz" (OuterVolumeSpecName: "kube-api-access-vfrdz") pod "abb4c773-a284-4c5e-8f93-4b023241adaf" (UID: "abb4c773-a284-4c5e-8f93-4b023241adaf"). InnerVolumeSpecName "kube-api-access-vfrdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.312861 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abb4c773-a284-4c5e-8f93-4b023241adaf" (UID: "abb4c773-a284-4c5e-8f93-4b023241adaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.321554 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data" (OuterVolumeSpecName: "config-data") pod "abb4c773-a284-4c5e-8f93-4b023241adaf" (UID: "abb4c773-a284-4c5e-8f93-4b023241adaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.376989 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abb4c773-a284-4c5e-8f93-4b023241adaf-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.377229 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.377242 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb4c773-a284-4c5e-8f93-4b023241adaf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.377250 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfrdz\" (UniqueName: \"kubernetes.io/projected/abb4c773-a284-4c5e-8f93-4b023241adaf-kube-api-access-vfrdz\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.543181 5002 generic.go:334] "Generic (PLEG): container finished" podID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerID="79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958" exitCode=0 Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.543291 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerDied","Data":"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958"} Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.543322 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"abb4c773-a284-4c5e-8f93-4b023241adaf","Type":"ContainerDied","Data":"6f024419f653b7ab40bc587bad7b8ef7aa41a25ed061d8b8603ff423f5e63607"} Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.543342 5002 scope.go:117] "RemoveContainer" containerID="79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.543505 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.550576 5002 generic.go:334] "Generic (PLEG): container finished" podID="41221834-182a-406d-9b98-26a51af2f6d7" containerID="20c4859c26269158c99e23fa5b9c3ed79fd6b49fda3e81fb2b16b381d28aa7f3" exitCode=0 Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.550618 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerDied","Data":"20c4859c26269158c99e23fa5b9c3ed79fd6b49fda3e81fb2b16b381d28aa7f3"} Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.576971 5002 scope.go:117] "RemoveContainer" containerID="f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.590898 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.599375 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.609462 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:22 crc kubenswrapper[5002]: E1014 08:10:22.609940 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-api" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.609962 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-api" Oct 14 08:10:22 crc kubenswrapper[5002]: E1014 08:10:22.610003 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-log" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.610012 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-log" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.610231 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-log" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.610266 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" containerName="nova-api-api" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.611481 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.617000 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.617262 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.618759 5002 scope.go:117] "RemoveContainer" containerID="79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958" Oct 14 08:10:22 crc kubenswrapper[5002]: E1014 08:10:22.619220 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958\": container with ID starting with 79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958 not found: ID does not exist" containerID="79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.619311 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958"} err="failed to get container status \"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958\": rpc error: code = NotFound desc = could not find container \"79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958\": container with ID starting with 79b8593fe8b754d35927c6028dc8cf7e13658f723b7c929fd9f3ae8e8484e958 not found: ID does not exist" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.619402 5002 scope.go:117] "RemoveContainer" containerID="f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.619620 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 08:10:22 crc kubenswrapper[5002]: E1014 08:10:22.619732 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d\": container with ID starting with f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d not found: ID does not exist" containerID="f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.619763 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d"} err="failed to get container status \"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d\": rpc error: code = NotFound desc = could not find container \"f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d\": container with ID starting with f7b242025bc70e5266babf3a9bb4875f9b5414399b1ac68cf2dfff01ed83814d not found: ID does not exist" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.627041 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.656039 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682475 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682545 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682584 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqmw4\" (UniqueName: \"kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682699 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682714 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682735 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682794 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.682859 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data\") pod \"41221834-182a-406d-9b98-26a51af2f6d7\" (UID: \"41221834-182a-406d-9b98-26a51af2f6d7\") " Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683095 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683115 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683144 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683180 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg848\" (UniqueName: \"kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683219 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683247 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.683873 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.686181 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.701133 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4" (OuterVolumeSpecName: "kube-api-access-hqmw4") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "kube-api-access-hqmw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.701701 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts" (OuterVolumeSpecName: "scripts") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.730214 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.774197 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.784791 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.784830 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785500 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785603 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg848\" (UniqueName: \"kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785695 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785751 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785923 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785937 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785947 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqmw4\" (UniqueName: \"kubernetes.io/projected/41221834-182a-406d-9b98-26a51af2f6d7-kube-api-access-hqmw4\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785957 5002 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785965 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.785973 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41221834-182a-406d-9b98-26a51af2f6d7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.786003 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.790527 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.790957 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.790883 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.791630 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.793183 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.803746 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg848\" (UniqueName: \"kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848\") pod \"nova-api-0\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " pod="openstack/nova-api-0" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.818468 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data" (OuterVolumeSpecName: "config-data") pod "41221834-182a-406d-9b98-26a51af2f6d7" (UID: "41221834-182a-406d-9b98-26a51af2f6d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.886996 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.887142 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41221834-182a-406d-9b98-26a51af2f6d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:22 crc kubenswrapper[5002]: I1014 08:10:22.955741 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.403335 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.566573 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerStarted","Data":"01003db914779eae8f3f6542be0b9282523ecbaee816d0f1b215cede92df3909"} Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.573137 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41221834-182a-406d-9b98-26a51af2f6d7","Type":"ContainerDied","Data":"63f36b6de4c610fd955703cbd349d03741f5759ad55b52b043f3cf55936572ed"} Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.573200 5002 scope.go:117] "RemoveContainer" containerID="504bd83fd70888b4b45dee5862c8c70ec0a3e8c9ea3905bd60a47dcf35231316" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.573332 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.600457 5002 scope.go:117] "RemoveContainer" containerID="64ddbdc17f84c52b6793ce6454c85d26ad30594b15e9be7d21e5c4239ca52252" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.622998 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.636152 5002 scope.go:117] "RemoveContainer" containerID="20c4859c26269158c99e23fa5b9c3ed79fd6b49fda3e81fb2b16b381d28aa7f3" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.640971 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.652630 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:23 crc kubenswrapper[5002]: E1014 08:10:23.653194 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-central-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653222 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-central-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: E1014 08:10:23.653239 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-notification-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653249 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-notification-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: E1014 08:10:23.653265 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="sg-core" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653274 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="sg-core" Oct 14 08:10:23 crc kubenswrapper[5002]: E1014 08:10:23.653299 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="proxy-httpd" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653307 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="proxy-httpd" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653530 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-central-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653550 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="ceilometer-notification-agent" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653568 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="sg-core" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.653593 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="41221834-182a-406d-9b98-26a51af2f6d7" containerName="proxy-httpd" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.655716 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.657697 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.658039 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.658379 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.658462 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.683933 5002 scope.go:117] "RemoveContainer" containerID="5d18d0e57fd977b82f23214d9fda6d136e76756a0c149dacc804f2f81af2952e" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700642 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700698 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700728 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700767 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700868 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700923 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvcwf\" (UniqueName: \"kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.700978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.701024 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.733345 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41221834-182a-406d-9b98-26a51af2f6d7" path="/var/lib/kubelet/pods/41221834-182a-406d-9b98-26a51af2f6d7/volumes" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.734311 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb4c773-a284-4c5e-8f93-4b023241adaf" path="/var/lib/kubelet/pods/abb4c773-a284-4c5e-8f93-4b023241adaf/volumes" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.802956 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803018 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803081 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803103 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803123 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803151 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803212 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.803237 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvcwf\" (UniqueName: \"kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.804385 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.804598 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.808055 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.808055 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.809367 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.811816 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.814263 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.829583 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvcwf\" (UniqueName: \"kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf\") pod \"ceilometer-0\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " pod="openstack/ceilometer-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.856175 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.884052 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:23 crc kubenswrapper[5002]: I1014 08:10:23.982226 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.474546 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:10:24 crc kubenswrapper[5002]: W1014 08:10:24.485298 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5a8ceb1_0d12_4bdb_b903_b99c044d8a4a.slice/crio-11531cdd2629e2633cb4e436d819a692a48d5c590d0f65370f9431905f3fb52a WatchSource:0}: Error finding container 11531cdd2629e2633cb4e436d819a692a48d5c590d0f65370f9431905f3fb52a: Status 404 returned error can't find the container with id 11531cdd2629e2633cb4e436d819a692a48d5c590d0f65370f9431905f3fb52a Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.584660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerStarted","Data":"11531cdd2629e2633cb4e436d819a692a48d5c590d0f65370f9431905f3fb52a"} Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.587269 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerStarted","Data":"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea"} Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.587306 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerStarted","Data":"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506"} Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.614752 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.614733445 podStartE2EDuration="2.614733445s" podCreationTimestamp="2025-10-14 08:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:24.608057151 +0000 UTC m=+1157.589296613" watchObservedRunningTime="2025-10-14 08:10:24.614733445 +0000 UTC m=+1157.595972897" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.618115 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.847516 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ccnf8"] Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.849291 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.851342 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.856154 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ccnf8"] Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.858263 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.935908 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.936065 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.936157 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:24 crc kubenswrapper[5002]: I1014 08:10:24.936215 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvfv\" (UniqueName: \"kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.037476 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.037596 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.037683 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.037721 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvfv\" (UniqueName: \"kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.045851 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.046366 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.050167 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.053280 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvfv\" (UniqueName: \"kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv\") pod \"nova-cell1-cell-mapping-ccnf8\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.172371 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.601257 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerStarted","Data":"bbef91002f42dce8d3ae557c3f9842e0589749f929357b4acdb773f81c4373a0"} Oct 14 08:10:25 crc kubenswrapper[5002]: I1014 08:10:25.694876 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ccnf8"] Oct 14 08:10:25 crc kubenswrapper[5002]: W1014 08:10:25.705075 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7749993e_625e_4787_9d58_ebdc54edf429.slice/crio-34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34 WatchSource:0}: Error finding container 34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34: Status 404 returned error can't find the container with id 34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34 Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.064998 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.143462 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.148424 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="dnsmasq-dns" containerID="cri-o://d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8" gracePeriod=10 Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.597635 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.613560 5002 generic.go:334] "Generic (PLEG): container finished" podID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerID="d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8" exitCode=0 Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.613616 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" event={"ID":"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4","Type":"ContainerDied","Data":"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8"} Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.613642 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" event={"ID":"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4","Type":"ContainerDied","Data":"daff9bed9e51f47ee64fb18093936f5ae86b6f166598b88a5900da4a21a931d5"} Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.613661 5002 scope.go:117] "RemoveContainer" containerID="d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.613764 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8c9888b9-q7xlp" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.632990 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerStarted","Data":"beabcf21650b91cde42a85e773d3aaa2388b081e7251ca8d8d6fc0d3ac407829"} Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.639392 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ccnf8" event={"ID":"7749993e-625e-4787-9d58-ebdc54edf429","Type":"ContainerStarted","Data":"8784cbc047e8110b041c0f8dc89f0ae2571b6943832959a68cc90e0224393cde"} Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.639442 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ccnf8" event={"ID":"7749993e-625e-4787-9d58-ebdc54edf429","Type":"ContainerStarted","Data":"34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34"} Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.646185 5002 scope.go:117] "RemoveContainer" containerID="0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.665721 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ccnf8" podStartSLOduration=2.665696642 podStartE2EDuration="2.665696642s" podCreationTimestamp="2025-10-14 08:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:26.65437742 +0000 UTC m=+1159.635616882" watchObservedRunningTime="2025-10-14 08:10:26.665696642 +0000 UTC m=+1159.646936114" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.675052 5002 scope.go:117] "RemoveContainer" containerID="d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8" Oct 14 08:10:26 crc kubenswrapper[5002]: E1014 08:10:26.676206 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8\": container with ID starting with d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8 not found: ID does not exist" containerID="d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.676239 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8"} err="failed to get container status \"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8\": rpc error: code = NotFound desc = could not find container \"d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8\": container with ID starting with d0d715e525e9ff7b9af486a645bdd23531d8e92d7d689e1cf8895f6a5e2633c8 not found: ID does not exist" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.676261 5002 scope.go:117] "RemoveContainer" containerID="0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57" Oct 14 08:10:26 crc kubenswrapper[5002]: E1014 08:10:26.676669 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57\": container with ID starting with 0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57 not found: ID does not exist" containerID="0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.676726 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57"} err="failed to get container status \"0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57\": rpc error: code = NotFound desc = could not find container \"0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57\": container with ID starting with 0e4738f2424c07bb45c2f3894131347ff049cabce5a53cace9a03d16d25bec57 not found: ID does not exist" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.689628 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb\") pod \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.689787 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb\") pod \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.689880 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config\") pod \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.689908 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc\") pod \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.690021 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76kwz\" (UniqueName: \"kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz\") pod \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\" (UID: \"d5f69148-aa22-4f4f-a42e-ebe0205bc7b4\") " Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.696179 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz" (OuterVolumeSpecName: "kube-api-access-76kwz") pod "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" (UID: "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4"). InnerVolumeSpecName "kube-api-access-76kwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.736547 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config" (OuterVolumeSpecName: "config") pod "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" (UID: "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.737498 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" (UID: "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.754117 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" (UID: "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.774369 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" (UID: "d5f69148-aa22-4f4f-a42e-ebe0205bc7b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.792954 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.792985 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.792996 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.793004 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:26 crc kubenswrapper[5002]: I1014 08:10:26.793014 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76kwz\" (UniqueName: \"kubernetes.io/projected/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4-kube-api-access-76kwz\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:27 crc kubenswrapper[5002]: I1014 08:10:27.075698 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:10:27 crc kubenswrapper[5002]: I1014 08:10:27.084406 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d8c9888b9-q7xlp"] Oct 14 08:10:27 crc kubenswrapper[5002]: I1014 08:10:27.651240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerStarted","Data":"6d30cf1060a60054748dff0eb336d3fbebc717dccd47808ca981fe76d3a77992"} Oct 14 08:10:27 crc kubenswrapper[5002]: I1014 08:10:27.753151 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" path="/var/lib/kubelet/pods/d5f69148-aa22-4f4f-a42e-ebe0205bc7b4/volumes" Oct 14 08:10:29 crc kubenswrapper[5002]: I1014 08:10:29.682737 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerStarted","Data":"53c4ee8860f08e35c23df2d3535b4380e7996a01e9b4483f0fb50f76f349a3f5"} Oct 14 08:10:29 crc kubenswrapper[5002]: I1014 08:10:29.684688 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:10:29 crc kubenswrapper[5002]: I1014 08:10:29.726054 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5269297120000003 podStartE2EDuration="6.72603463s" podCreationTimestamp="2025-10-14 08:10:23 +0000 UTC" firstStartedPulling="2025-10-14 08:10:24.48811336 +0000 UTC m=+1157.469352812" lastFinishedPulling="2025-10-14 08:10:28.687218278 +0000 UTC m=+1161.668457730" observedRunningTime="2025-10-14 08:10:29.713183295 +0000 UTC m=+1162.694422767" watchObservedRunningTime="2025-10-14 08:10:29.72603463 +0000 UTC m=+1162.707274092" Oct 14 08:10:30 crc kubenswrapper[5002]: I1014 08:10:30.696329 5002 generic.go:334] "Generic (PLEG): container finished" podID="7749993e-625e-4787-9d58-ebdc54edf429" containerID="8784cbc047e8110b041c0f8dc89f0ae2571b6943832959a68cc90e0224393cde" exitCode=0 Oct 14 08:10:30 crc kubenswrapper[5002]: I1014 08:10:30.696416 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ccnf8" event={"ID":"7749993e-625e-4787-9d58-ebdc54edf429","Type":"ContainerDied","Data":"8784cbc047e8110b041c0f8dc89f0ae2571b6943832959a68cc90e0224393cde"} Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.093927 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.214012 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data\") pod \"7749993e-625e-4787-9d58-ebdc54edf429\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.214459 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle\") pod \"7749993e-625e-4787-9d58-ebdc54edf429\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.214494 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts\") pod \"7749993e-625e-4787-9d58-ebdc54edf429\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.214578 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvfv\" (UniqueName: \"kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv\") pod \"7749993e-625e-4787-9d58-ebdc54edf429\" (UID: \"7749993e-625e-4787-9d58-ebdc54edf429\") " Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.220670 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv" (OuterVolumeSpecName: "kube-api-access-8rvfv") pod "7749993e-625e-4787-9d58-ebdc54edf429" (UID: "7749993e-625e-4787-9d58-ebdc54edf429"). InnerVolumeSpecName "kube-api-access-8rvfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.221807 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts" (OuterVolumeSpecName: "scripts") pod "7749993e-625e-4787-9d58-ebdc54edf429" (UID: "7749993e-625e-4787-9d58-ebdc54edf429"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.240392 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data" (OuterVolumeSpecName: "config-data") pod "7749993e-625e-4787-9d58-ebdc54edf429" (UID: "7749993e-625e-4787-9d58-ebdc54edf429"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.274833 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7749993e-625e-4787-9d58-ebdc54edf429" (UID: "7749993e-625e-4787-9d58-ebdc54edf429"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.317125 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.317404 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.317421 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7749993e-625e-4787-9d58-ebdc54edf429-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.317450 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvfv\" (UniqueName: \"kubernetes.io/projected/7749993e-625e-4787-9d58-ebdc54edf429-kube-api-access-8rvfv\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.719138 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ccnf8" event={"ID":"7749993e-625e-4787-9d58-ebdc54edf429","Type":"ContainerDied","Data":"34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34"} Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.719188 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34bd652c72aec15a4d1c9e28474854321d2e212c22ad8bdb6f6fd384f4b23f34" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.719255 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ccnf8" Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.932244 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.932602 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-log" containerID="cri-o://a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" gracePeriod=30 Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.933289 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-api" containerID="cri-o://e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" gracePeriod=30 Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.943330 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:32 crc kubenswrapper[5002]: I1014 08:10:32.943594 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6e0662d7-dc68-4c29-a61a-64b03c15057b" containerName="nova-scheduler-scheduler" containerID="cri-o://b846ebb53cb1875c506c96afc2d86693fc36142011ea7fd373ba8eef43afa869" gracePeriod=30 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.016334 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.018125 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" containerID="cri-o://6589b31335c69bd2566d93b8a7ec5e5cf35be148c7597bfca9a9c50992feda9e" gracePeriod=30 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.018236 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" containerID="cri-o://58e93834475a6a8014a5b86627c2f7fc7e4429dbe8060f057453badbb7ded9ba" gracePeriod=30 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.536100 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647386 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647461 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647527 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647590 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg848\" (UniqueName: \"kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647672 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.647718 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data\") pod \"c3d81834-c787-45ea-964a-2937d69c5fb7\" (UID: \"c3d81834-c787-45ea-964a-2937d69c5fb7\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.650187 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs" (OuterVolumeSpecName: "logs") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.671129 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848" (OuterVolumeSpecName: "kube-api-access-sg848") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "kube-api-access-sg848". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.747565 5002 generic.go:334] "Generic (PLEG): container finished" podID="6e0662d7-dc68-4c29-a61a-64b03c15057b" containerID="b846ebb53cb1875c506c96afc2d86693fc36142011ea7fd373ba8eef43afa869" exitCode=0 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.749875 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3d81834-c787-45ea-964a-2937d69c5fb7-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.749897 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg848\" (UniqueName: \"kubernetes.io/projected/c3d81834-c787-45ea-964a-2937d69c5fb7-kube-api-access-sg848\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.751711 5002 generic.go:334] "Generic (PLEG): container finished" podID="26720557-b8fa-4460-bf5f-1922335bb86b" containerID="6589b31335c69bd2566d93b8a7ec5e5cf35be148c7597bfca9a9c50992feda9e" exitCode=143 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754305 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6e0662d7-dc68-4c29-a61a-64b03c15057b","Type":"ContainerDied","Data":"b846ebb53cb1875c506c96afc2d86693fc36142011ea7fd373ba8eef43afa869"} Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754340 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerDied","Data":"6589b31335c69bd2566d93b8a7ec5e5cf35be148c7597bfca9a9c50992feda9e"} Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754492 5002 generic.go:334] "Generic (PLEG): container finished" podID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerID="e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" exitCode=0 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754508 5002 generic.go:334] "Generic (PLEG): container finished" podID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerID="a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" exitCode=143 Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754521 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerDied","Data":"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506"} Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754533 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerDied","Data":"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea"} Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754542 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c3d81834-c787-45ea-964a-2937d69c5fb7","Type":"ContainerDied","Data":"01003db914779eae8f3f6542be0b9282523ecbaee816d0f1b215cede92df3909"} Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754558 5002 scope.go:117] "RemoveContainer" containerID="e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.754587 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.760489 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data" (OuterVolumeSpecName: "config-data") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.764185 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.777025 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.777844 5002 scope.go:117] "RemoveContainer" containerID="a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.788655 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.795416 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c3d81834-c787-45ea-964a-2937d69c5fb7" (UID: "c3d81834-c787-45ea-964a-2937d69c5fb7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.817637 5002 scope.go:117] "RemoveContainer" containerID="e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" Oct 14 08:10:33 crc kubenswrapper[5002]: E1014 08:10:33.818254 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506\": container with ID starting with e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506 not found: ID does not exist" containerID="e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.818295 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506"} err="failed to get container status \"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506\": rpc error: code = NotFound desc = could not find container \"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506\": container with ID starting with e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506 not found: ID does not exist" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.818320 5002 scope.go:117] "RemoveContainer" containerID="a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" Oct 14 08:10:33 crc kubenswrapper[5002]: E1014 08:10:33.818547 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea\": container with ID starting with a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea not found: ID does not exist" containerID="a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.818573 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea"} err="failed to get container status \"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea\": rpc error: code = NotFound desc = could not find container \"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea\": container with ID starting with a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea not found: ID does not exist" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.818594 5002 scope.go:117] "RemoveContainer" containerID="e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.818999 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506"} err="failed to get container status \"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506\": rpc error: code = NotFound desc = could not find container \"e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506\": container with ID starting with e1f7cd03d15f3baaf69921fa33c39dda532a693b66209bef0cacb8c85ab89506 not found: ID does not exist" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.819025 5002 scope.go:117] "RemoveContainer" containerID="a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.819232 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea"} err="failed to get container status \"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea\": rpc error: code = NotFound desc = could not find container \"a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea\": container with ID starting with a4c695c200b4d279dcbc39f0bf84860657094de0bb6ca7841cab0ad91f70dfea not found: ID does not exist" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.851373 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data\") pod \"6e0662d7-dc68-4c29-a61a-64b03c15057b\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.851550 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kbh8\" (UniqueName: \"kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8\") pod \"6e0662d7-dc68-4c29-a61a-64b03c15057b\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.851651 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle\") pod \"6e0662d7-dc68-4c29-a61a-64b03c15057b\" (UID: \"6e0662d7-dc68-4c29-a61a-64b03c15057b\") " Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.854796 5002 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.854812 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8" (OuterVolumeSpecName: "kube-api-access-7kbh8") pod "6e0662d7-dc68-4c29-a61a-64b03c15057b" (UID: "6e0662d7-dc68-4c29-a61a-64b03c15057b"). InnerVolumeSpecName "kube-api-access-7kbh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.854822 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.854861 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.854872 5002 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3d81834-c787-45ea-964a-2937d69c5fb7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.889296 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e0662d7-dc68-4c29-a61a-64b03c15057b" (UID: "6e0662d7-dc68-4c29-a61a-64b03c15057b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.894016 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data" (OuterVolumeSpecName: "config-data") pod "6e0662d7-dc68-4c29-a61a-64b03c15057b" (UID: "6e0662d7-dc68-4c29-a61a-64b03c15057b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.956511 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.956545 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kbh8\" (UniqueName: \"kubernetes.io/projected/6e0662d7-dc68-4c29-a61a-64b03c15057b-kube-api-access-7kbh8\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:33 crc kubenswrapper[5002]: I1014 08:10:33.956557 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0662d7-dc68-4c29-a61a-64b03c15057b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.131515 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.147277 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155011 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155517 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-log" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155542 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-log" Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155567 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7749993e-625e-4787-9d58-ebdc54edf429" containerName="nova-manage" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155575 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7749993e-625e-4787-9d58-ebdc54edf429" containerName="nova-manage" Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155612 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="dnsmasq-dns" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155622 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="dnsmasq-dns" Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155630 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0662d7-dc68-4c29-a61a-64b03c15057b" containerName="nova-scheduler-scheduler" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155638 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0662d7-dc68-4c29-a61a-64b03c15057b" containerName="nova-scheduler-scheduler" Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155656 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-api" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155665 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-api" Oct 14 08:10:34 crc kubenswrapper[5002]: E1014 08:10:34.155681 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="init" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155689 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="init" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155952 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7749993e-625e-4787-9d58-ebdc54edf429" containerName="nova-manage" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.155986 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-api" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.156003 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f69148-aa22-4f4f-a42e-ebe0205bc7b4" containerName="dnsmasq-dns" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.156014 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0662d7-dc68-4c29-a61a-64b03c15057b" containerName="nova-scheduler-scheduler" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.156024 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" containerName="nova-api-log" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.157321 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.159933 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.160144 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.160321 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.163191 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.261674 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-config-data\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.261729 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.261796 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.261922 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j5kc\" (UniqueName: \"kubernetes.io/projected/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-kube-api-access-4j5kc\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.261953 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-logs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.262322 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.364681 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j5kc\" (UniqueName: \"kubernetes.io/projected/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-kube-api-access-4j5kc\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.364780 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-logs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.365029 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.365150 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-config-data\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.365254 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.365321 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.365733 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-logs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.372279 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-config-data\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.373234 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.374941 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.376597 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.384900 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j5kc\" (UniqueName: \"kubernetes.io/projected/55d153d2-2f4c-4e77-9a7a-fa120ac94da2-kube-api-access-4j5kc\") pod \"nova-api-0\" (UID: \"55d153d2-2f4c-4e77-9a7a-fa120ac94da2\") " pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.485247 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.767284 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6e0662d7-dc68-4c29-a61a-64b03c15057b","Type":"ContainerDied","Data":"bcdaf945492057679227b1cc7a41e8cce3eb868d7725e33b5403a6f7a2eae40a"} Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.767351 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.767529 5002 scope.go:117] "RemoveContainer" containerID="b846ebb53cb1875c506c96afc2d86693fc36142011ea7fd373ba8eef43afa869" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.810952 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.823912 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.832913 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.834377 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.838337 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.844633 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.877576 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.877636 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dc7s\" (UniqueName: \"kubernetes.io/projected/59facc52-a7a5-4a65-b7ba-67e9d8267de2-kube-api-access-9dc7s\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.877824 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-config-data\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.939642 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.980109 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.980488 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dc7s\" (UniqueName: \"kubernetes.io/projected/59facc52-a7a5-4a65-b7ba-67e9d8267de2-kube-api-access-9dc7s\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.980591 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-config-data\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.985182 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-config-data\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.988360 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59facc52-a7a5-4a65-b7ba-67e9d8267de2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:34 crc kubenswrapper[5002]: I1014 08:10:34.995903 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dc7s\" (UniqueName: \"kubernetes.io/projected/59facc52-a7a5-4a65-b7ba-67e9d8267de2-kube-api-access-9dc7s\") pod \"nova-scheduler-0\" (UID: \"59facc52-a7a5-4a65-b7ba-67e9d8267de2\") " pod="openstack/nova-scheduler-0" Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.192110 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:10:35 crc kubenswrapper[5002]: W1014 08:10:35.466689 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59facc52_a7a5_4a65_b7ba_67e9d8267de2.slice/crio-396bba49644bee8ec5ef83334a3fbb2ee39261635e38dcbd91b063578ef46bd6 WatchSource:0}: Error finding container 396bba49644bee8ec5ef83334a3fbb2ee39261635e38dcbd91b063578ef46bd6: Status 404 returned error can't find the container with id 396bba49644bee8ec5ef83334a3fbb2ee39261635e38dcbd91b063578ef46bd6 Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.467094 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.732631 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0662d7-dc68-4c29-a61a-64b03c15057b" path="/var/lib/kubelet/pods/6e0662d7-dc68-4c29-a61a-64b03c15057b/volumes" Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.733752 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3d81834-c787-45ea-964a-2937d69c5fb7" path="/var/lib/kubelet/pods/c3d81834-c787-45ea-964a-2937d69c5fb7/volumes" Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.780659 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"59facc52-a7a5-4a65-b7ba-67e9d8267de2","Type":"ContainerStarted","Data":"992d2457fa88d648b96755ec7592df1aebbd4ce46a5fa6198922413957590a53"} Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.780710 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"59facc52-a7a5-4a65-b7ba-67e9d8267de2","Type":"ContainerStarted","Data":"396bba49644bee8ec5ef83334a3fbb2ee39261635e38dcbd91b063578ef46bd6"} Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.786168 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55d153d2-2f4c-4e77-9a7a-fa120ac94da2","Type":"ContainerStarted","Data":"b6a6b1080cda1ae3a7b2561a0389b375ddaf6170ff14f643cd503e45717e04a7"} Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.786208 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55d153d2-2f4c-4e77-9a7a-fa120ac94da2","Type":"ContainerStarted","Data":"b8dac4cd659623471ebfa891fbe0b15fa48e0f176c2692bceb6efb88f13130cd"} Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.786221 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55d153d2-2f4c-4e77-9a7a-fa120ac94da2","Type":"ContainerStarted","Data":"50ccbe83f398b79b0b5dceec71e6662da05a32792d39621358f3b5b787623fda"} Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.815757 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.8157353889999999 podStartE2EDuration="1.815735389s" podCreationTimestamp="2025-10-14 08:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:35.804939762 +0000 UTC m=+1168.786179224" watchObservedRunningTime="2025-10-14 08:10:35.815735389 +0000 UTC m=+1168.796974851" Oct 14 08:10:35 crc kubenswrapper[5002]: I1014 08:10:35.831653 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.831626178 podStartE2EDuration="1.831626178s" podCreationTimestamp="2025-10-14 08:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:35.828263145 +0000 UTC m=+1168.809502637" watchObservedRunningTime="2025-10-14 08:10:35.831626178 +0000 UTC m=+1168.812865640" Oct 14 08:10:36 crc kubenswrapper[5002]: I1014 08:10:36.438932 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": read tcp 10.217.0.2:53874->10.217.0.185:8775: read: connection reset by peer" Oct 14 08:10:36 crc kubenswrapper[5002]: I1014 08:10:36.439028 5002 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.185:8775/\": read tcp 10.217.0.2:53866->10.217.0.185:8775: read: connection reset by peer" Oct 14 08:10:36 crc kubenswrapper[5002]: I1014 08:10:36.796161 5002 generic.go:334] "Generic (PLEG): container finished" podID="26720557-b8fa-4460-bf5f-1922335bb86b" containerID="58e93834475a6a8014a5b86627c2f7fc7e4429dbe8060f057453badbb7ded9ba" exitCode=0 Oct 14 08:10:36 crc kubenswrapper[5002]: I1014 08:10:36.796274 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerDied","Data":"58e93834475a6a8014a5b86627c2f7fc7e4429dbe8060f057453badbb7ded9ba"} Oct 14 08:10:36 crc kubenswrapper[5002]: I1014 08:10:36.947951 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.031948 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data\") pod \"26720557-b8fa-4460-bf5f-1922335bb86b\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.032025 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs\") pod \"26720557-b8fa-4460-bf5f-1922335bb86b\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.032069 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st257\" (UniqueName: \"kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257\") pod \"26720557-b8fa-4460-bf5f-1922335bb86b\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.032132 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle\") pod \"26720557-b8fa-4460-bf5f-1922335bb86b\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.032183 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs\") pod \"26720557-b8fa-4460-bf5f-1922335bb86b\" (UID: \"26720557-b8fa-4460-bf5f-1922335bb86b\") " Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.032851 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs" (OuterVolumeSpecName: "logs") pod "26720557-b8fa-4460-bf5f-1922335bb86b" (UID: "26720557-b8fa-4460-bf5f-1922335bb86b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.047218 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257" (OuterVolumeSpecName: "kube-api-access-st257") pod "26720557-b8fa-4460-bf5f-1922335bb86b" (UID: "26720557-b8fa-4460-bf5f-1922335bb86b"). InnerVolumeSpecName "kube-api-access-st257". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.066130 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data" (OuterVolumeSpecName: "config-data") pod "26720557-b8fa-4460-bf5f-1922335bb86b" (UID: "26720557-b8fa-4460-bf5f-1922335bb86b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.068591 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26720557-b8fa-4460-bf5f-1922335bb86b" (UID: "26720557-b8fa-4460-bf5f-1922335bb86b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.104972 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "26720557-b8fa-4460-bf5f-1922335bb86b" (UID: "26720557-b8fa-4460-bf5f-1922335bb86b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.138000 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.138293 5002 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.138370 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st257\" (UniqueName: \"kubernetes.io/projected/26720557-b8fa-4460-bf5f-1922335bb86b-kube-api-access-st257\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.138446 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26720557-b8fa-4460-bf5f-1922335bb86b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.138508 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26720557-b8fa-4460-bf5f-1922335bb86b-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.809879 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"26720557-b8fa-4460-bf5f-1922335bb86b","Type":"ContainerDied","Data":"9d40e60617616e1150d8b8fb388e1836aa71ddcb2c8aed5b566e486cc5f5a668"} Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.810034 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.810217 5002 scope.go:117] "RemoveContainer" containerID="58e93834475a6a8014a5b86627c2f7fc7e4429dbe8060f057453badbb7ded9ba" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.846548 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.860677 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.872190 5002 scope.go:117] "RemoveContainer" containerID="6589b31335c69bd2566d93b8a7ec5e5cf35be148c7597bfca9a9c50992feda9e" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.873055 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:37 crc kubenswrapper[5002]: E1014 08:10:37.873425 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.873441 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" Oct 14 08:10:37 crc kubenswrapper[5002]: E1014 08:10:37.873468 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.873475 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.873639 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-metadata" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.873664 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" containerName="nova-metadata-log" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.874635 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.877109 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.877254 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.885313 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.951307 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-logs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.951356 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-config-data\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.951411 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.951578 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:37 crc kubenswrapper[5002]: I1014 08:10:37.951760 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7cl\" (UniqueName: \"kubernetes.io/projected/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-kube-api-access-bv7cl\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.052915 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7cl\" (UniqueName: \"kubernetes.io/projected/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-kube-api-access-bv7cl\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.053026 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-logs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.053049 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-config-data\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.053080 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.053106 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.054108 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-logs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.057644 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.057716 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.059194 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-config-data\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.073667 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7cl\" (UniqueName: \"kubernetes.io/projected/e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5-kube-api-access-bv7cl\") pod \"nova-metadata-0\" (UID: \"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5\") " pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.224554 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.728129 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:10:38 crc kubenswrapper[5002]: I1014 08:10:38.828271 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5","Type":"ContainerStarted","Data":"879dc19e15ba3e1a64ab068aef519cea99e88b37e92b8f7802c60f4e4cb9c00c"} Oct 14 08:10:39 crc kubenswrapper[5002]: I1014 08:10:39.739954 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26720557-b8fa-4460-bf5f-1922335bb86b" path="/var/lib/kubelet/pods/26720557-b8fa-4460-bf5f-1922335bb86b/volumes" Oct 14 08:10:39 crc kubenswrapper[5002]: I1014 08:10:39.856530 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5","Type":"ContainerStarted","Data":"07a120643012d11f1ab3b9754d168402734b6f5d5abb4b0efbbfaec8df6ffffa"} Oct 14 08:10:39 crc kubenswrapper[5002]: I1014 08:10:39.856589 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5","Type":"ContainerStarted","Data":"bea98ce641d95740bc84174dc1532d78838ce7147cdaacbf619179edb4eb188b"} Oct 14 08:10:39 crc kubenswrapper[5002]: I1014 08:10:39.894415 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.894386342 podStartE2EDuration="2.894386342s" podCreationTimestamp="2025-10-14 08:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:10:39.890120974 +0000 UTC m=+1172.871360496" watchObservedRunningTime="2025-10-14 08:10:39.894386342 +0000 UTC m=+1172.875625824" Oct 14 08:10:40 crc kubenswrapper[5002]: I1014 08:10:40.193506 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:10:43 crc kubenswrapper[5002]: I1014 08:10:43.224826 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:10:43 crc kubenswrapper[5002]: I1014 08:10:43.225507 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:10:44 crc kubenswrapper[5002]: I1014 08:10:44.487127 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:10:44 crc kubenswrapper[5002]: I1014 08:10:44.487207 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 08:10:45 crc kubenswrapper[5002]: I1014 08:10:45.193090 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 08:10:45 crc kubenswrapper[5002]: I1014 08:10:45.226825 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 08:10:45 crc kubenswrapper[5002]: I1014 08:10:45.507147 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55d153d2-2f4c-4e77-9a7a-fa120ac94da2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:45 crc kubenswrapper[5002]: I1014 08:10:45.507113 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55d153d2-2f4c-4e77-9a7a-fa120ac94da2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:45 crc kubenswrapper[5002]: I1014 08:10:45.994026 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 08:10:48 crc kubenswrapper[5002]: I1014 08:10:48.225154 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:10:48 crc kubenswrapper[5002]: I1014 08:10:48.225600 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 08:10:49 crc kubenswrapper[5002]: I1014 08:10:49.242080 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:49 crc kubenswrapper[5002]: I1014 08:10:49.242080 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 08:10:53 crc kubenswrapper[5002]: I1014 08:10:53.998828 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:10:54 crc kubenswrapper[5002]: I1014 08:10:54.493824 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:10:54 crc kubenswrapper[5002]: I1014 08:10:54.495149 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:10:54 crc kubenswrapper[5002]: I1014 08:10:54.500404 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:10:54 crc kubenswrapper[5002]: I1014 08:10:54.514310 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 08:10:55 crc kubenswrapper[5002]: I1014 08:10:55.043947 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 08:10:55 crc kubenswrapper[5002]: I1014 08:10:55.050756 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 08:10:58 crc kubenswrapper[5002]: I1014 08:10:58.232948 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:10:58 crc kubenswrapper[5002]: I1014 08:10:58.236609 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 08:10:58 crc kubenswrapper[5002]: I1014 08:10:58.240148 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:10:59 crc kubenswrapper[5002]: I1014 08:10:59.111802 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 08:11:07 crc kubenswrapper[5002]: I1014 08:11:07.199018 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:08 crc kubenswrapper[5002]: I1014 08:11:08.209912 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:11 crc kubenswrapper[5002]: I1014 08:11:11.784321 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="rabbitmq" containerID="cri-o://9efc43d1e74ae820fb9f25e41c46693c27be9380501b3fde947346eda7a623f4" gracePeriod=604796 Oct 14 08:11:12 crc kubenswrapper[5002]: I1014 08:11:12.669523 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="rabbitmq" containerID="cri-o://1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be" gracePeriod=604796 Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.333923 5002 generic.go:334] "Generic (PLEG): container finished" podID="f7fe047a-78a9-4abf-a948-167ba509729c" containerID="9efc43d1e74ae820fb9f25e41c46693c27be9380501b3fde947346eda7a623f4" exitCode=0 Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.334303 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerDied","Data":"9efc43d1e74ae820fb9f25e41c46693c27be9380501b3fde947346eda7a623f4"} Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.493282 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.536964 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537013 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537129 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537213 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsz8\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537258 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537334 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537355 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537380 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537422 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537448 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.537479 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info\") pod \"f7fe047a-78a9-4abf-a948-167ba509729c\" (UID: \"f7fe047a-78a9-4abf-a948-167ba509729c\") " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.538293 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.542992 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.543183 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.545399 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.546273 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.546737 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.546891 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info" (OuterVolumeSpecName: "pod-info") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.547134 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8" (OuterVolumeSpecName: "kube-api-access-dbsz8") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "kube-api-access-dbsz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.586860 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data" (OuterVolumeSpecName: "config-data") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.599353 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf" (OuterVolumeSpecName: "server-conf") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639296 5002 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639564 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsz8\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-kube-api-access-dbsz8\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639632 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639686 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639741 5002 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe047a-78a9-4abf-a948-167ba509729c-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639801 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639869 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639931 5002 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe047a-78a9-4abf-a948-167ba509729c-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.639991 5002 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe047a-78a9-4abf-a948-167ba509729c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.640071 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.660680 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.667028 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f7fe047a-78a9-4abf-a948-167ba509729c" (UID: "f7fe047a-78a9-4abf-a948-167ba509729c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.742036 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:18 crc kubenswrapper[5002]: I1014 08:11:18.742065 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe047a-78a9-4abf-a948-167ba509729c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.265324 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.345370 5002 generic.go:334] "Generic (PLEG): container finished" podID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerID="1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be" exitCode=0 Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.345501 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.345798 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerDied","Data":"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be"} Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.345855 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b16df05d-7859-4d03-8a4b-ca31e68d1d32","Type":"ContainerDied","Data":"44d459f7ed92756cbebddc17016393b9d510953fefcd538061bff7d3d4269d07"} Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.345875 5002 scope.go:117] "RemoveContainer" containerID="1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355625 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355668 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355799 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355881 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355911 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.355943 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356000 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356033 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356092 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356138 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356166 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hmj4\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4\") pod \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\" (UID: \"b16df05d-7859-4d03-8a4b-ca31e68d1d32\") " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356454 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356606 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356911 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.356930 5002 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.357298 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.357404 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f7fe047a-78a9-4abf-a948-167ba509729c","Type":"ContainerDied","Data":"0cb81a0212979b5d88f7ff7c929cd5f4ae717239fdb60a73e486b89f050ab982"} Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.357515 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.363321 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.367509 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4" (OuterVolumeSpecName: "kube-api-access-5hmj4") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "kube-api-access-5hmj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.372176 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.372584 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.376021 5002 scope.go:117] "RemoveContainer" containerID="1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.387248 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info" (OuterVolumeSpecName: "pod-info") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.401888 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data" (OuterVolumeSpecName: "config-data") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.450446 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf" (OuterVolumeSpecName: "server-conf") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458801 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458847 5002 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b16df05d-7859-4d03-8a4b-ca31e68d1d32-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458861 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hmj4\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-kube-api-access-5hmj4\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458875 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458886 5002 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b16df05d-7859-4d03-8a4b-ca31e68d1d32-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458897 5002 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b16df05d-7859-4d03-8a4b-ca31e68d1d32-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458923 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.458934 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.490826 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.493038 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.494905 5002 scope.go:117] "RemoveContainer" containerID="1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be" Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.495506 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be\": container with ID starting with 1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be not found: ID does not exist" containerID="1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.495552 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be"} err="failed to get container status \"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be\": rpc error: code = NotFound desc = could not find container \"1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be\": container with ID starting with 1d24da761a6c870c028e3ef8ff8ecb84d2944a9e71739812fab5c1dc1a7f27be not found: ID does not exist" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.495613 5002 scope.go:117] "RemoveContainer" containerID="1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9" Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.496157 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9\": container with ID starting with 1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9 not found: ID does not exist" containerID="1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.496262 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9"} err="failed to get container status \"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9\": rpc error: code = NotFound desc = could not find container \"1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9\": container with ID starting with 1f3326b427bb63a1bcfa2a628727cef94e37a31a97147fdfc94a6c6f6f0ab8d9 not found: ID does not exist" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.496353 5002 scope.go:117] "RemoveContainer" containerID="9efc43d1e74ae820fb9f25e41c46693c27be9380501b3fde947346eda7a623f4" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.508723 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.518866 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.519374 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="setup-container" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.519440 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="setup-container" Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.519509 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.519649 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.519722 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="setup-container" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.519774 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="setup-container" Oct 14 08:11:19 crc kubenswrapper[5002]: E1014 08:11:19.519832 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.519912 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.520147 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.520244 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" containerName="rabbitmq" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.521289 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.523261 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b16df05d-7859-4d03-8a4b-ca31e68d1d32" (UID: "b16df05d-7859-4d03-8a4b-ca31e68d1d32"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.527876 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tgsb8" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.527921 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.528202 5002 scope.go:117] "RemoveContainer" containerID="2336f75047f1dbec2b388320d890aaa908a5b9cf671f7d3dd3b7b2a45675dbaa" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.528696 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.528822 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.529565 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.529685 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.530313 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.540955 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560586 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6d6v\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-kube-api-access-n6d6v\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560627 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560673 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcf9ff36-3590-4122-881c-6331338dd63f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560693 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcf9ff36-3590-4122-881c-6331338dd63f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560731 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560763 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-config-data\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560782 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560809 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.560831 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.561604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.561651 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.561719 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.562367 5002 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b16df05d-7859-4d03-8a4b-ca31e68d1d32-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663783 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6d6v\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-kube-api-access-n6d6v\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663829 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663886 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcf9ff36-3590-4122-881c-6331338dd63f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663906 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcf9ff36-3590-4122-881c-6331338dd63f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663946 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663981 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-config-data\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.663998 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664028 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664052 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664092 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664125 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664599 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664694 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664736 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664805 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.665743 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.664876 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bcf9ff36-3590-4122-881c-6331338dd63f-config-data\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.669685 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcf9ff36-3590-4122-881c-6331338dd63f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.669698 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.670088 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcf9ff36-3590-4122-881c-6331338dd63f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.675775 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.680261 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.681171 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6d6v\" (UniqueName: \"kubernetes.io/projected/bcf9ff36-3590-4122-881c-6331338dd63f-kube-api-access-n6d6v\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.687510 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.701851 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"bcf9ff36-3590-4122-881c-6331338dd63f\") " pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.705092 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.706557 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.708737 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.709146 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.709303 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.709484 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-z9926" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.709598 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.711025 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.711166 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.743764 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b16df05d-7859-4d03-8a4b-ca31e68d1d32" path="/var/lib/kubelet/pods/b16df05d-7859-4d03-8a4b-ca31e68d1d32/volumes" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.744517 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7fe047a-78a9-4abf-a948-167ba509729c" path="/var/lib/kubelet/pods/f7fe047a-78a9-4abf-a948-167ba509729c/volumes" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.745006 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766024 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766316 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94673e4e-a0de-4a23-981d-1a6c7ea055ff-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766438 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766554 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766652 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766814 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.766955 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.767056 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.767155 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.767279 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5nht\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-kube-api-access-x5nht\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.767410 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94673e4e-a0de-4a23-981d-1a6c7ea055ff-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.843749 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868621 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94673e4e-a0de-4a23-981d-1a6c7ea055ff-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868682 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868718 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868739 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868805 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868863 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868891 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868912 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868943 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5nht\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-kube-api-access-x5nht\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.868986 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94673e4e-a0de-4a23-981d-1a6c7ea055ff-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.869020 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.869471 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.870065 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.870095 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.871012 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.871570 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.872557 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94673e4e-a0de-4a23-981d-1a6c7ea055ff-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.874799 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94673e4e-a0de-4a23-981d-1a6c7ea055ff-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.875941 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94673e4e-a0de-4a23-981d-1a6c7ea055ff-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.878933 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.882751 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.894049 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5nht\" (UniqueName: \"kubernetes.io/projected/94673e4e-a0de-4a23-981d-1a6c7ea055ff-kube-api-access-x5nht\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:19 crc kubenswrapper[5002]: I1014 08:11:19.922631 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94673e4e-a0de-4a23-981d-1a6c7ea055ff\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:20 crc kubenswrapper[5002]: I1014 08:11:20.033952 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:20 crc kubenswrapper[5002]: I1014 08:11:20.334036 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:11:20 crc kubenswrapper[5002]: I1014 08:11:20.368811 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bcf9ff36-3590-4122-881c-6331338dd63f","Type":"ContainerStarted","Data":"d45e73afa7e5b5711ad1005f0fc1aa2e60441178d0f1006cd23923e05159077c"} Oct 14 08:11:20 crc kubenswrapper[5002]: I1014 08:11:20.522916 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:11:20 crc kubenswrapper[5002]: W1014 08:11:20.533394 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94673e4e_a0de_4a23_981d_1a6c7ea055ff.slice/crio-96cbf43a74131d139134fcec6d0c76c03dc71f3eb14af4ad1404537af9c5a71e WatchSource:0}: Error finding container 96cbf43a74131d139134fcec6d0c76c03dc71f3eb14af4ad1404537af9c5a71e: Status 404 returned error can't find the container with id 96cbf43a74131d139134fcec6d0c76c03dc71f3eb14af4ad1404537af9c5a71e Oct 14 08:11:21 crc kubenswrapper[5002]: I1014 08:11:21.386482 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94673e4e-a0de-4a23-981d-1a6c7ea055ff","Type":"ContainerStarted","Data":"96cbf43a74131d139134fcec6d0c76c03dc71f3eb14af4ad1404537af9c5a71e"} Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.816370 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.818498 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.820449 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.843281 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931206 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931293 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931346 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931381 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931563 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95p56\" (UniqueName: \"kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:22 crc kubenswrapper[5002]: I1014 08:11:22.931618 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.032859 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.032950 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95p56\" (UniqueName: \"kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.032979 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.033053 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.033116 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.033159 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.033819 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.034133 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.034354 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.034415 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.034761 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.053500 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95p56\" (UniqueName: \"kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56\") pod \"dnsmasq-dns-588d944c9f-5nm8t\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.170138 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.414089 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94673e4e-a0de-4a23-981d-1a6c7ea055ff","Type":"ContainerStarted","Data":"1a11868d8f52864336ea0ba033709fd40963235fc5bf6a28e98b2f5ac5c60ab0"} Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.416312 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bcf9ff36-3590-4122-881c-6331338dd63f","Type":"ContainerStarted","Data":"44e7985ed6cc64d44c528e7a6f876f927d34c665d4b33ef24e4845691c3086a9"} Oct 14 08:11:23 crc kubenswrapper[5002]: I1014 08:11:23.676705 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:24 crc kubenswrapper[5002]: I1014 08:11:24.440457 5002 generic.go:334] "Generic (PLEG): container finished" podID="01df8306-c096-4160-a9c5-b9e5253036d9" containerID="f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa" exitCode=0 Oct 14 08:11:24 crc kubenswrapper[5002]: I1014 08:11:24.441291 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" event={"ID":"01df8306-c096-4160-a9c5-b9e5253036d9","Type":"ContainerDied","Data":"f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa"} Oct 14 08:11:24 crc kubenswrapper[5002]: I1014 08:11:24.441414 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" event={"ID":"01df8306-c096-4160-a9c5-b9e5253036d9","Type":"ContainerStarted","Data":"44b80db0593b151df44826c739ef9d59331cf086502e240a8e80b939649b4fed"} Oct 14 08:11:25 crc kubenswrapper[5002]: I1014 08:11:25.453830 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" event={"ID":"01df8306-c096-4160-a9c5-b9e5253036d9","Type":"ContainerStarted","Data":"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291"} Oct 14 08:11:25 crc kubenswrapper[5002]: I1014 08:11:25.454068 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:25 crc kubenswrapper[5002]: I1014 08:11:25.480069 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" podStartSLOduration=3.480051922 podStartE2EDuration="3.480051922s" podCreationTimestamp="2025-10-14 08:11:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:11:25.476292558 +0000 UTC m=+1218.457532080" watchObservedRunningTime="2025-10-14 08:11:25.480051922 +0000 UTC m=+1218.461291374" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.172455 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.236617 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.236860 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="dnsmasq-dns" containerID="cri-o://63115e709943833ad05e7ec6407a64046c52be7d78a17239489c100bd50817af" gracePeriod=10 Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.383308 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.396665 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.433131 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.476696 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.476812 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.476842 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.476891 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.476945 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgfpd\" (UniqueName: \"kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.477005 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.552201 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5891d56-c935-449d-a120-14ebe283db57" containerID="63115e709943833ad05e7ec6407a64046c52be7d78a17239489c100bd50817af" exitCode=0 Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.552267 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" event={"ID":"c5891d56-c935-449d-a120-14ebe283db57","Type":"ContainerDied","Data":"63115e709943833ad05e7ec6407a64046c52be7d78a17239489c100bd50817af"} Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.578787 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.578865 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.578922 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.578946 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.578974 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgfpd\" (UniqueName: \"kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.579008 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.579794 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.580558 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.580632 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.580696 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.580733 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.609662 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgfpd\" (UniqueName: \"kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd\") pod \"dnsmasq-dns-5c8c578c57-x8rz8\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.759130 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.825834 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.991725 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb\") pod \"c5891d56-c935-449d-a120-14ebe283db57\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.991795 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4fkk\" (UniqueName: \"kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk\") pod \"c5891d56-c935-449d-a120-14ebe283db57\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.991899 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config\") pod \"c5891d56-c935-449d-a120-14ebe283db57\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.991999 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb\") pod \"c5891d56-c935-449d-a120-14ebe283db57\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.992080 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc\") pod \"c5891d56-c935-449d-a120-14ebe283db57\" (UID: \"c5891d56-c935-449d-a120-14ebe283db57\") " Oct 14 08:11:33 crc kubenswrapper[5002]: I1014 08:11:33.996386 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk" (OuterVolumeSpecName: "kube-api-access-v4fkk") pod "c5891d56-c935-449d-a120-14ebe283db57" (UID: "c5891d56-c935-449d-a120-14ebe283db57"). InnerVolumeSpecName "kube-api-access-v4fkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.041966 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config" (OuterVolumeSpecName: "config") pod "c5891d56-c935-449d-a120-14ebe283db57" (UID: "c5891d56-c935-449d-a120-14ebe283db57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.048321 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c5891d56-c935-449d-a120-14ebe283db57" (UID: "c5891d56-c935-449d-a120-14ebe283db57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.049678 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c5891d56-c935-449d-a120-14ebe283db57" (UID: "c5891d56-c935-449d-a120-14ebe283db57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.052883 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c5891d56-c935-449d-a120-14ebe283db57" (UID: "c5891d56-c935-449d-a120-14ebe283db57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.094461 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.094507 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.094524 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4fkk\" (UniqueName: \"kubernetes.io/projected/c5891d56-c935-449d-a120-14ebe283db57-kube-api-access-v4fkk\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.094536 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.094546 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5891d56-c935-449d-a120-14ebe283db57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.208283 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.565390 5002 generic.go:334] "Generic (PLEG): container finished" podID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerID="ae26742f397f08dbff20c2d52ada7f2514ff2daf722a6ae5d74e34105097687b" exitCode=0 Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.565567 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" event={"ID":"49cf58a5-bf2a-46bb-b08f-94e85686c2a6","Type":"ContainerDied","Data":"ae26742f397f08dbff20c2d52ada7f2514ff2daf722a6ae5d74e34105097687b"} Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.565676 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" event={"ID":"49cf58a5-bf2a-46bb-b08f-94e85686c2a6","Type":"ContainerStarted","Data":"f11c8cb3ee322722e199bbbf922869636a44157c7948e6c3f59663f5d44978c1"} Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.568321 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" event={"ID":"c5891d56-c935-449d-a120-14ebe283db57","Type":"ContainerDied","Data":"fb1dbea78883a1644d9bc2b0ff01e3f54d1ded8178ec3c28794a65d5c6ea7b9c"} Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.568382 5002 scope.go:117] "RemoveContainer" containerID="63115e709943833ad05e7ec6407a64046c52be7d78a17239489c100bd50817af" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.568529 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9cf55b5c-fpcpc" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.720897 5002 scope.go:117] "RemoveContainer" containerID="bb471d6a417baa5671dfb6daaf368b02bd70846e31713d8f0ce7cfd7697c3f55" Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.754334 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:11:34 crc kubenswrapper[5002]: I1014 08:11:34.762006 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f9cf55b5c-fpcpc"] Oct 14 08:11:35 crc kubenswrapper[5002]: I1014 08:11:35.582697 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" event={"ID":"49cf58a5-bf2a-46bb-b08f-94e85686c2a6","Type":"ContainerStarted","Data":"80769c05d7af0588e89ecda6c65d5f674afd365655adfc5a296051a624518034"} Oct 14 08:11:35 crc kubenswrapper[5002]: I1014 08:11:35.583053 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:35 crc kubenswrapper[5002]: I1014 08:11:35.605410 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" podStartSLOduration=2.6053968469999997 podStartE2EDuration="2.605396847s" podCreationTimestamp="2025-10-14 08:11:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:11:35.603013531 +0000 UTC m=+1228.584252993" watchObservedRunningTime="2025-10-14 08:11:35.605396847 +0000 UTC m=+1228.586636299" Oct 14 08:11:35 crc kubenswrapper[5002]: I1014 08:11:35.730935 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5891d56-c935-449d-a120-14ebe283db57" path="/var/lib/kubelet/pods/c5891d56-c935-449d-a120-14ebe283db57/volumes" Oct 14 08:11:39 crc kubenswrapper[5002]: I1014 08:11:39.217974 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:11:39 crc kubenswrapper[5002]: I1014 08:11:39.218441 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:11:43 crc kubenswrapper[5002]: I1014 08:11:43.762577 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:11:43 crc kubenswrapper[5002]: I1014 08:11:43.827143 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:43 crc kubenswrapper[5002]: I1014 08:11:43.827497 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="dnsmasq-dns" containerID="cri-o://7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291" gracePeriod=10 Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.357861 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.498760 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.498854 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.498984 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.499058 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95p56\" (UniqueName: \"kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.499123 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.499141 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc\") pod \"01df8306-c096-4160-a9c5-b9e5253036d9\" (UID: \"01df8306-c096-4160-a9c5-b9e5253036d9\") " Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.504292 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56" (OuterVolumeSpecName: "kube-api-access-95p56") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "kube-api-access-95p56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.545446 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.545551 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.563491 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.569073 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config" (OuterVolumeSpecName: "config") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.570138 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "01df8306-c096-4160-a9c5-b9e5253036d9" (UID: "01df8306-c096-4160-a9c5-b9e5253036d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601162 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601201 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601214 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601228 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601241 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/01df8306-c096-4160-a9c5-b9e5253036d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.601254 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95p56\" (UniqueName: \"kubernetes.io/projected/01df8306-c096-4160-a9c5-b9e5253036d9-kube-api-access-95p56\") on node \"crc\" DevicePath \"\"" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.690874 5002 generic.go:334] "Generic (PLEG): container finished" podID="01df8306-c096-4160-a9c5-b9e5253036d9" containerID="7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291" exitCode=0 Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.690942 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.690955 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" event={"ID":"01df8306-c096-4160-a9c5-b9e5253036d9","Type":"ContainerDied","Data":"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291"} Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.691032 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-588d944c9f-5nm8t" event={"ID":"01df8306-c096-4160-a9c5-b9e5253036d9","Type":"ContainerDied","Data":"44b80db0593b151df44826c739ef9d59331cf086502e240a8e80b939649b4fed"} Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.691071 5002 scope.go:117] "RemoveContainer" containerID="7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.723378 5002 scope.go:117] "RemoveContainer" containerID="f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.740441 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.746929 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-588d944c9f-5nm8t"] Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.763526 5002 scope.go:117] "RemoveContainer" containerID="7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291" Oct 14 08:11:44 crc kubenswrapper[5002]: E1014 08:11:44.764404 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291\": container with ID starting with 7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291 not found: ID does not exist" containerID="7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.764625 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291"} err="failed to get container status \"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291\": rpc error: code = NotFound desc = could not find container \"7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291\": container with ID starting with 7844a41cd36c69beb3a20200f7978cb072140b7790fcc0c5bae6a28b72b28291 not found: ID does not exist" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.764800 5002 scope.go:117] "RemoveContainer" containerID="f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa" Oct 14 08:11:44 crc kubenswrapper[5002]: E1014 08:11:44.765585 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa\": container with ID starting with f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa not found: ID does not exist" containerID="f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa" Oct 14 08:11:44 crc kubenswrapper[5002]: I1014 08:11:44.765658 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa"} err="failed to get container status \"f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa\": rpc error: code = NotFound desc = could not find container \"f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa\": container with ID starting with f2f0c0056f2691237c25305d074fbddbf63dd64de9a9063c168738d9013963fa not found: ID does not exist" Oct 14 08:11:45 crc kubenswrapper[5002]: I1014 08:11:45.739012 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" path="/var/lib/kubelet/pods/01df8306-c096-4160-a9c5-b9e5253036d9/volumes" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.060908 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6"] Oct 14 08:11:54 crc kubenswrapper[5002]: E1014 08:11:54.061918 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="init" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.061934 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="init" Oct 14 08:11:54 crc kubenswrapper[5002]: E1014 08:11:54.061953 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.061960 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: E1014 08:11:54.061994 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="init" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.062001 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="init" Oct 14 08:11:54 crc kubenswrapper[5002]: E1014 08:11:54.062021 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.062028 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.062199 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="01df8306-c096-4160-a9c5-b9e5253036d9" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.062227 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5891d56-c935-449d-a120-14ebe283db57" containerName="dnsmasq-dns" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.062935 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.064770 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.065438 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.065619 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.065520 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.092112 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6"] Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.205393 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.205448 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.205492 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.205588 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42rsk\" (UniqueName: \"kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.307735 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.307807 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.307873 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.307987 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42rsk\" (UniqueName: \"kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.313775 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.315939 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.316818 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.338638 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42rsk\" (UniqueName: \"kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:54 crc kubenswrapper[5002]: I1014 08:11:54.378108 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.023595 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6"] Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.036742 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.835701 5002 generic.go:334] "Generic (PLEG): container finished" podID="94673e4e-a0de-4a23-981d-1a6c7ea055ff" containerID="1a11868d8f52864336ea0ba033709fd40963235fc5bf6a28e98b2f5ac5c60ab0" exitCode=0 Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.835779 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94673e4e-a0de-4a23-981d-1a6c7ea055ff","Type":"ContainerDied","Data":"1a11868d8f52864336ea0ba033709fd40963235fc5bf6a28e98b2f5ac5c60ab0"} Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.844360 5002 generic.go:334] "Generic (PLEG): container finished" podID="bcf9ff36-3590-4122-881c-6331338dd63f" containerID="44e7985ed6cc64d44c528e7a6f876f927d34c665d4b33ef24e4845691c3086a9" exitCode=0 Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.844522 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bcf9ff36-3590-4122-881c-6331338dd63f","Type":"ContainerDied","Data":"44e7985ed6cc64d44c528e7a6f876f927d34c665d4b33ef24e4845691c3086a9"} Oct 14 08:11:55 crc kubenswrapper[5002]: I1014 08:11:55.850543 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" event={"ID":"be99665c-ce97-4ce5-8e47-2855db76123f","Type":"ContainerStarted","Data":"a37a459a22d18f61c3a7ddcf1b273a350de933fa6baf8a79804ef6475ddb92d1"} Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.861411 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"bcf9ff36-3590-4122-881c-6331338dd63f","Type":"ContainerStarted","Data":"c610da6396f7abae95c560d8598fde67332c9e524a5dcba8dd52656bd761ab7b"} Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.861912 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.866080 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94673e4e-a0de-4a23-981d-1a6c7ea055ff","Type":"ContainerStarted","Data":"b56acc89c7cf7612c5565341b5cfe8bbd3bd81f4bd2cbf6b61c01844f05e60f1"} Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.866261 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.889032 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.889013559 podStartE2EDuration="37.889013559s" podCreationTimestamp="2025-10-14 08:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:11:56.882618855 +0000 UTC m=+1249.863858337" watchObservedRunningTime="2025-10-14 08:11:56.889013559 +0000 UTC m=+1249.870253011" Oct 14 08:11:56 crc kubenswrapper[5002]: I1014 08:11:56.902740 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.902724159 podStartE2EDuration="37.902724159s" podCreationTimestamp="2025-10-14 08:11:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:11:56.901372023 +0000 UTC m=+1249.882611485" watchObservedRunningTime="2025-10-14 08:11:56.902724159 +0000 UTC m=+1249.883963611" Oct 14 08:12:04 crc kubenswrapper[5002]: I1014 08:12:04.956296 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" event={"ID":"be99665c-ce97-4ce5-8e47-2855db76123f","Type":"ContainerStarted","Data":"7c953250b86cad42d7e9267f8d5627e61dff0d0e304dcde3407698eb4bc111a1"} Oct 14 08:12:04 crc kubenswrapper[5002]: I1014 08:12:04.991369 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" podStartSLOduration=1.5005577209999998 podStartE2EDuration="10.991324642s" podCreationTimestamp="2025-10-14 08:11:54 +0000 UTC" firstStartedPulling="2025-10-14 08:11:55.036531845 +0000 UTC m=+1248.017771297" lastFinishedPulling="2025-10-14 08:12:04.527298726 +0000 UTC m=+1257.508538218" observedRunningTime="2025-10-14 08:12:04.973434688 +0000 UTC m=+1257.954674161" watchObservedRunningTime="2025-10-14 08:12:04.991324642 +0000 UTC m=+1257.972564124" Oct 14 08:12:09 crc kubenswrapper[5002]: I1014 08:12:09.218824 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:12:09 crc kubenswrapper[5002]: I1014 08:12:09.219473 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:12:09 crc kubenswrapper[5002]: I1014 08:12:09.848392 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:12:10 crc kubenswrapper[5002]: I1014 08:12:10.041499 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:12:17 crc kubenswrapper[5002]: I1014 08:12:17.089540 5002 generic.go:334] "Generic (PLEG): container finished" podID="be99665c-ce97-4ce5-8e47-2855db76123f" containerID="7c953250b86cad42d7e9267f8d5627e61dff0d0e304dcde3407698eb4bc111a1" exitCode=0 Oct 14 08:12:17 crc kubenswrapper[5002]: I1014 08:12:17.089647 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" event={"ID":"be99665c-ce97-4ce5-8e47-2855db76123f","Type":"ContainerDied","Data":"7c953250b86cad42d7e9267f8d5627e61dff0d0e304dcde3407698eb4bc111a1"} Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.669317 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.817468 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle\") pod \"be99665c-ce97-4ce5-8e47-2855db76123f\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.817537 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42rsk\" (UniqueName: \"kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk\") pod \"be99665c-ce97-4ce5-8e47-2855db76123f\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.817620 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key\") pod \"be99665c-ce97-4ce5-8e47-2855db76123f\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.817899 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory\") pod \"be99665c-ce97-4ce5-8e47-2855db76123f\" (UID: \"be99665c-ce97-4ce5-8e47-2855db76123f\") " Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.828075 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk" (OuterVolumeSpecName: "kube-api-access-42rsk") pod "be99665c-ce97-4ce5-8e47-2855db76123f" (UID: "be99665c-ce97-4ce5-8e47-2855db76123f"). InnerVolumeSpecName "kube-api-access-42rsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.828383 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "be99665c-ce97-4ce5-8e47-2855db76123f" (UID: "be99665c-ce97-4ce5-8e47-2855db76123f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.869941 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be99665c-ce97-4ce5-8e47-2855db76123f" (UID: "be99665c-ce97-4ce5-8e47-2855db76123f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.876157 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory" (OuterVolumeSpecName: "inventory") pod "be99665c-ce97-4ce5-8e47-2855db76123f" (UID: "be99665c-ce97-4ce5-8e47-2855db76123f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.921402 5002 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.921503 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42rsk\" (UniqueName: \"kubernetes.io/projected/be99665c-ce97-4ce5-8e47-2855db76123f-kube-api-access-42rsk\") on node \"crc\" DevicePath \"\"" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.921517 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:12:18 crc kubenswrapper[5002]: I1014 08:12:18.921530 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be99665c-ce97-4ce5-8e47-2855db76123f-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.118493 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" event={"ID":"be99665c-ce97-4ce5-8e47-2855db76123f","Type":"ContainerDied","Data":"a37a459a22d18f61c3a7ddcf1b273a350de933fa6baf8a79804ef6475ddb92d1"} Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.118556 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a37a459a22d18f61c3a7ddcf1b273a350de933fa6baf8a79804ef6475ddb92d1" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.119295 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.208856 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt"] Oct 14 08:12:19 crc kubenswrapper[5002]: E1014 08:12:19.209394 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be99665c-ce97-4ce5-8e47-2855db76123f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.209426 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="be99665c-ce97-4ce5-8e47-2855db76123f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.209697 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="be99665c-ce97-4ce5-8e47-2855db76123f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.210772 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.214196 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.214347 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.214467 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.214585 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.226091 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt"] Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.328361 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.328421 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wdnn\" (UniqueName: \"kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.328464 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.328551 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.429789 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wdnn\" (UniqueName: \"kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.429881 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.430008 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.430098 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.434668 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.436636 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.436940 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.455249 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wdnn\" (UniqueName: \"kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.540191 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:12:19 crc kubenswrapper[5002]: I1014 08:12:19.845415 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt"] Oct 14 08:12:19 crc kubenswrapper[5002]: W1014 08:12:19.849623 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5f087b0_c7ff_46d2_8134_94f0879ff23c.slice/crio-7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509 WatchSource:0}: Error finding container 7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509: Status 404 returned error can't find the container with id 7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509 Oct 14 08:12:20 crc kubenswrapper[5002]: I1014 08:12:20.131043 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" event={"ID":"e5f087b0-c7ff-46d2-8134-94f0879ff23c","Type":"ContainerStarted","Data":"7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509"} Oct 14 08:12:21 crc kubenswrapper[5002]: I1014 08:12:21.145994 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" event={"ID":"e5f087b0-c7ff-46d2-8134-94f0879ff23c","Type":"ContainerStarted","Data":"3371c2d98c7c31e767da693319499d2d3528e35a8c95219bf4615f157bdd9c81"} Oct 14 08:12:21 crc kubenswrapper[5002]: I1014 08:12:21.169661 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" podStartSLOduration=1.3598048280000001 podStartE2EDuration="2.169640023s" podCreationTimestamp="2025-10-14 08:12:19 +0000 UTC" firstStartedPulling="2025-10-14 08:12:19.854547649 +0000 UTC m=+1272.835787101" lastFinishedPulling="2025-10-14 08:12:20.664382804 +0000 UTC m=+1273.645622296" observedRunningTime="2025-10-14 08:12:21.165589515 +0000 UTC m=+1274.146828987" watchObservedRunningTime="2025-10-14 08:12:21.169640023 +0000 UTC m=+1274.150879485" Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.218983 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.219603 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.219662 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.221085 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.221227 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5" gracePeriod=600 Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.366353 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5" exitCode=0 Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.366418 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5"} Oct 14 08:12:39 crc kubenswrapper[5002]: I1014 08:12:39.366476 5002 scope.go:117] "RemoveContainer" containerID="17d1ae774546942ebdb1bb3cd4135c6302ccf330af268b804dd17d7ffd47533b" Oct 14 08:12:40 crc kubenswrapper[5002]: I1014 08:12:40.384483 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63"} Oct 14 08:13:21 crc kubenswrapper[5002]: I1014 08:13:21.683631 5002 scope.go:117] "RemoveContainer" containerID="5c6cde16a988f75021402f33ef371619f677ef4c8da9bcb8df53c1c0b1bf63c2" Oct 14 08:13:21 crc kubenswrapper[5002]: I1014 08:13:21.715562 5002 scope.go:117] "RemoveContainer" containerID="6288dc12ae757ad9e3b98e3c0c7ec5840048d2e57bcc9f727475073969ded676" Oct 14 08:14:21 crc kubenswrapper[5002]: I1014 08:14:21.859239 5002 scope.go:117] "RemoveContainer" containerID="46badaaa8912c6af3248be56592226d71d3ebc94ebd469fac2d66d7e695c6261" Oct 14 08:14:21 crc kubenswrapper[5002]: I1014 08:14:21.895113 5002 scope.go:117] "RemoveContainer" containerID="d928716ed2af799033fe6b4c73237e741f2698d058fb0a12896fb0628a0ddd20" Oct 14 08:14:21 crc kubenswrapper[5002]: I1014 08:14:21.920267 5002 scope.go:117] "RemoveContainer" containerID="ac854a13512a9d841b051aa8b108fe2f0839603d57044323c1936182255189ec" Oct 14 08:14:39 crc kubenswrapper[5002]: I1014 08:14:39.218776 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:14:39 crc kubenswrapper[5002]: I1014 08:14:39.219549 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.475728 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.478283 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.510457 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.595633 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm8hr\" (UniqueName: \"kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.596112 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.596379 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.698294 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm8hr\" (UniqueName: \"kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.698384 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.698450 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.699132 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.699140 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.731975 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm8hr\" (UniqueName: \"kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr\") pod \"redhat-marketplace-fncbd\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:44 crc kubenswrapper[5002]: I1014 08:14:44.812649 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:45 crc kubenswrapper[5002]: I1014 08:14:45.292744 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:45 crc kubenswrapper[5002]: W1014 08:14:45.304400 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20798be0_f280_4bd8_b43e_7684e2f1b700.slice/crio-9f260fc248fc3d5a18edb534a06502c383accefa6a6f46d1dce6045b4da0df79 WatchSource:0}: Error finding container 9f260fc248fc3d5a18edb534a06502c383accefa6a6f46d1dce6045b4da0df79: Status 404 returned error can't find the container with id 9f260fc248fc3d5a18edb534a06502c383accefa6a6f46d1dce6045b4da0df79 Oct 14 08:14:45 crc kubenswrapper[5002]: I1014 08:14:45.927411 5002 generic.go:334] "Generic (PLEG): container finished" podID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerID="25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1" exitCode=0 Oct 14 08:14:45 crc kubenswrapper[5002]: I1014 08:14:45.927515 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerDied","Data":"25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1"} Oct 14 08:14:45 crc kubenswrapper[5002]: I1014 08:14:45.927700 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerStarted","Data":"9f260fc248fc3d5a18edb534a06502c383accefa6a6f46d1dce6045b4da0df79"} Oct 14 08:14:46 crc kubenswrapper[5002]: I1014 08:14:46.939852 5002 generic.go:334] "Generic (PLEG): container finished" podID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerID="d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2" exitCode=0 Oct 14 08:14:46 crc kubenswrapper[5002]: I1014 08:14:46.939963 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerDied","Data":"d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2"} Oct 14 08:14:47 crc kubenswrapper[5002]: I1014 08:14:47.954497 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerStarted","Data":"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521"} Oct 14 08:14:47 crc kubenswrapper[5002]: I1014 08:14:47.986243 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fncbd" podStartSLOduration=2.411212907 podStartE2EDuration="3.986210757s" podCreationTimestamp="2025-10-14 08:14:44 +0000 UTC" firstStartedPulling="2025-10-14 08:14:45.929358619 +0000 UTC m=+1418.910598071" lastFinishedPulling="2025-10-14 08:14:47.504356429 +0000 UTC m=+1420.485595921" observedRunningTime="2025-10-14 08:14:47.973589566 +0000 UTC m=+1420.954829058" watchObservedRunningTime="2025-10-14 08:14:47.986210757 +0000 UTC m=+1420.967450229" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.762193 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.766080 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.778429 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.883546 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-477z5\" (UniqueName: \"kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.883886 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.884104 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.986190 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-477z5\" (UniqueName: \"kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.986315 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.986520 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.987205 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:53 crc kubenswrapper[5002]: I1014 08:14:53.987341 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:54 crc kubenswrapper[5002]: I1014 08:14:54.015856 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-477z5\" (UniqueName: \"kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5\") pod \"certified-operators-lwsdx\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:54 crc kubenswrapper[5002]: I1014 08:14:54.104010 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:14:54 crc kubenswrapper[5002]: I1014 08:14:54.812751 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:54 crc kubenswrapper[5002]: I1014 08:14:54.813104 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:54 crc kubenswrapper[5002]: I1014 08:14:54.895897 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:55 crc kubenswrapper[5002]: I1014 08:14:55.077550 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:55 crc kubenswrapper[5002]: I1014 08:14:55.152573 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:14:56 crc kubenswrapper[5002]: I1014 08:14:56.037198 5002 generic.go:334] "Generic (PLEG): container finished" podID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerID="ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3" exitCode=0 Oct 14 08:14:56 crc kubenswrapper[5002]: I1014 08:14:56.037286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerDied","Data":"ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3"} Oct 14 08:14:56 crc kubenswrapper[5002]: I1014 08:14:56.037663 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerStarted","Data":"0accd0436d78e890cb81ecb8128d140fee883dfbf6fbf9ea4a89661f26960be7"} Oct 14 08:14:56 crc kubenswrapper[5002]: I1014 08:14:56.918162 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.053218 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerStarted","Data":"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1"} Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.053432 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fncbd" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="registry-server" containerID="cri-o://e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521" gracePeriod=2 Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.594638 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.663981 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities\") pod \"20798be0-f280-4bd8-b43e-7684e2f1b700\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.664109 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm8hr\" (UniqueName: \"kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr\") pod \"20798be0-f280-4bd8-b43e-7684e2f1b700\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.664133 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content\") pod \"20798be0-f280-4bd8-b43e-7684e2f1b700\" (UID: \"20798be0-f280-4bd8-b43e-7684e2f1b700\") " Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.668752 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities" (OuterVolumeSpecName: "utilities") pod "20798be0-f280-4bd8-b43e-7684e2f1b700" (UID: "20798be0-f280-4bd8-b43e-7684e2f1b700"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.680249 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr" (OuterVolumeSpecName: "kube-api-access-mm8hr") pod "20798be0-f280-4bd8-b43e-7684e2f1b700" (UID: "20798be0-f280-4bd8-b43e-7684e2f1b700"). InnerVolumeSpecName "kube-api-access-mm8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.687196 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20798be0-f280-4bd8-b43e-7684e2f1b700" (UID: "20798be0-f280-4bd8-b43e-7684e2f1b700"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.766726 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.766779 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm8hr\" (UniqueName: \"kubernetes.io/projected/20798be0-f280-4bd8-b43e-7684e2f1b700-kube-api-access-mm8hr\") on node \"crc\" DevicePath \"\"" Oct 14 08:14:57 crc kubenswrapper[5002]: I1014 08:14:57.766801 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20798be0-f280-4bd8-b43e-7684e2f1b700-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.067215 5002 generic.go:334] "Generic (PLEG): container finished" podID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerID="d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1" exitCode=0 Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.067316 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerDied","Data":"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1"} Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.075440 5002 generic.go:334] "Generic (PLEG): container finished" podID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerID="e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521" exitCode=0 Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.075488 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerDied","Data":"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521"} Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.075529 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fncbd" event={"ID":"20798be0-f280-4bd8-b43e-7684e2f1b700","Type":"ContainerDied","Data":"9f260fc248fc3d5a18edb534a06502c383accefa6a6f46d1dce6045b4da0df79"} Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.075558 5002 scope.go:117] "RemoveContainer" containerID="e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.075745 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fncbd" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.122919 5002 scope.go:117] "RemoveContainer" containerID="d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.131430 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.144056 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fncbd"] Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.167325 5002 scope.go:117] "RemoveContainer" containerID="25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.201568 5002 scope.go:117] "RemoveContainer" containerID="e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521" Oct 14 08:14:58 crc kubenswrapper[5002]: E1014 08:14:58.202112 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521\": container with ID starting with e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521 not found: ID does not exist" containerID="e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.202182 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521"} err="failed to get container status \"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521\": rpc error: code = NotFound desc = could not find container \"e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521\": container with ID starting with e6d6db396b16ebd00f44f93fc0b31b7c7a833a7963e15c5b5329bd582f7bb521 not found: ID does not exist" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.202214 5002 scope.go:117] "RemoveContainer" containerID="d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2" Oct 14 08:14:58 crc kubenswrapper[5002]: E1014 08:14:58.202541 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2\": container with ID starting with d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2 not found: ID does not exist" containerID="d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.202571 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2"} err="failed to get container status \"d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2\": rpc error: code = NotFound desc = could not find container \"d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2\": container with ID starting with d8533b2a6a4718f54c7b643c6b278b413f5ed9f0fb186613b5efee68731515a2 not found: ID does not exist" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.202592 5002 scope.go:117] "RemoveContainer" containerID="25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1" Oct 14 08:14:58 crc kubenswrapper[5002]: E1014 08:14:58.202895 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1\": container with ID starting with 25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1 not found: ID does not exist" containerID="25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1" Oct 14 08:14:58 crc kubenswrapper[5002]: I1014 08:14:58.202926 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1"} err="failed to get container status \"25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1\": rpc error: code = NotFound desc = could not find container \"25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1\": container with ID starting with 25ea98a59338e5ff5de6af15a8701f47644d5a0fbe91ed9b9d5a73dae9d4b5b1 not found: ID does not exist" Oct 14 08:14:59 crc kubenswrapper[5002]: I1014 08:14:59.086991 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerStarted","Data":"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671"} Oct 14 08:14:59 crc kubenswrapper[5002]: I1014 08:14:59.105219 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lwsdx" podStartSLOduration=3.609285609 podStartE2EDuration="6.105167168s" podCreationTimestamp="2025-10-14 08:14:53 +0000 UTC" firstStartedPulling="2025-10-14 08:14:56.039284599 +0000 UTC m=+1429.020524051" lastFinishedPulling="2025-10-14 08:14:58.535166148 +0000 UTC m=+1431.516405610" observedRunningTime="2025-10-14 08:14:59.102420634 +0000 UTC m=+1432.083660096" watchObservedRunningTime="2025-10-14 08:14:59.105167168 +0000 UTC m=+1432.086406620" Oct 14 08:14:59 crc kubenswrapper[5002]: I1014 08:14:59.735269 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" path="/var/lib/kubelet/pods/20798be0-f280-4bd8-b43e-7684e2f1b700/volumes" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.168405 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh"] Oct 14 08:15:00 crc kubenswrapper[5002]: E1014 08:15:00.170502 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="extract-utilities" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.170632 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="extract-utilities" Oct 14 08:15:00 crc kubenswrapper[5002]: E1014 08:15:00.170779 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="extract-content" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.171216 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="extract-content" Oct 14 08:15:00 crc kubenswrapper[5002]: E1014 08:15:00.171340 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.171440 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.171889 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="20798be0-f280-4bd8-b43e-7684e2f1b700" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.173109 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.176086 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.181271 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.186037 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh"] Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.216293 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsmx\" (UniqueName: \"kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.216694 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.216805 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.318749 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.318958 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.319067 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsmx\" (UniqueName: \"kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.321354 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.332481 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.333755 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsmx\" (UniqueName: \"kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx\") pod \"collect-profiles-29340495-25mdh\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.494976 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:00 crc kubenswrapper[5002]: I1014 08:15:00.801880 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh"] Oct 14 08:15:01 crc kubenswrapper[5002]: I1014 08:15:01.112980 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" event={"ID":"4efd0fd5-8a8f-401f-a14f-ea0dee615274","Type":"ContainerStarted","Data":"f76d702845733377cb3c900878a6f406025144e810740b65c715deeb7c43e8d6"} Oct 14 08:15:01 crc kubenswrapper[5002]: I1014 08:15:01.113293 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" event={"ID":"4efd0fd5-8a8f-401f-a14f-ea0dee615274","Type":"ContainerStarted","Data":"90d9f224b4f46d9627fa884abb6b1b65e0516ce11bcfdf1de434ad6027096a46"} Oct 14 08:15:01 crc kubenswrapper[5002]: I1014 08:15:01.138855 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" podStartSLOduration=1.138812829 podStartE2EDuration="1.138812829s" podCreationTimestamp="2025-10-14 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:15:01.130518345 +0000 UTC m=+1434.111757817" watchObservedRunningTime="2025-10-14 08:15:01.138812829 +0000 UTC m=+1434.120052291" Oct 14 08:15:02 crc kubenswrapper[5002]: I1014 08:15:02.128495 5002 generic.go:334] "Generic (PLEG): container finished" podID="4efd0fd5-8a8f-401f-a14f-ea0dee615274" containerID="f76d702845733377cb3c900878a6f406025144e810740b65c715deeb7c43e8d6" exitCode=0 Oct 14 08:15:02 crc kubenswrapper[5002]: I1014 08:15:02.128559 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" event={"ID":"4efd0fd5-8a8f-401f-a14f-ea0dee615274","Type":"ContainerDied","Data":"f76d702845733377cb3c900878a6f406025144e810740b65c715deeb7c43e8d6"} Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.584132 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.692517 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume\") pod \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.692684 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbsmx\" (UniqueName: \"kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx\") pod \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.692730 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume\") pod \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\" (UID: \"4efd0fd5-8a8f-401f-a14f-ea0dee615274\") " Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.695000 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume" (OuterVolumeSpecName: "config-volume") pod "4efd0fd5-8a8f-401f-a14f-ea0dee615274" (UID: "4efd0fd5-8a8f-401f-a14f-ea0dee615274"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.699967 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4efd0fd5-8a8f-401f-a14f-ea0dee615274" (UID: "4efd0fd5-8a8f-401f-a14f-ea0dee615274"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.700024 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx" (OuterVolumeSpecName: "kube-api-access-bbsmx") pod "4efd0fd5-8a8f-401f-a14f-ea0dee615274" (UID: "4efd0fd5-8a8f-401f-a14f-ea0dee615274"). InnerVolumeSpecName "kube-api-access-bbsmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.795115 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4efd0fd5-8a8f-401f-a14f-ea0dee615274-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.795153 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbsmx\" (UniqueName: \"kubernetes.io/projected/4efd0fd5-8a8f-401f-a14f-ea0dee615274-kube-api-access-bbsmx\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[5002]: I1014 08:15:03.795163 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4efd0fd5-8a8f-401f-a14f-ea0dee615274-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.105108 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.105672 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.156224 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.156262 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh" event={"ID":"4efd0fd5-8a8f-401f-a14f-ea0dee615274","Type":"ContainerDied","Data":"90d9f224b4f46d9627fa884abb6b1b65e0516ce11bcfdf1de434ad6027096a46"} Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.156315 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90d9f224b4f46d9627fa884abb6b1b65e0516ce11bcfdf1de434ad6027096a46" Oct 14 08:15:04 crc kubenswrapper[5002]: I1014 08:15:04.164614 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:05 crc kubenswrapper[5002]: I1014 08:15:05.229727 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:05 crc kubenswrapper[5002]: I1014 08:15:05.292492 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.186019 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lwsdx" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="registry-server" containerID="cri-o://4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671" gracePeriod=2 Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.740807 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.780482 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities\") pod \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.780528 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content\") pod \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.780742 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-477z5\" (UniqueName: \"kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5\") pod \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\" (UID: \"e5dbefe8-13da-4839-a4b2-b8864bb698a2\") " Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.785116 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities" (OuterVolumeSpecName: "utilities") pod "e5dbefe8-13da-4839-a4b2-b8864bb698a2" (UID: "e5dbefe8-13da-4839-a4b2-b8864bb698a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.791985 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5" (OuterVolumeSpecName: "kube-api-access-477z5") pod "e5dbefe8-13da-4839-a4b2-b8864bb698a2" (UID: "e5dbefe8-13da-4839-a4b2-b8864bb698a2"). InnerVolumeSpecName "kube-api-access-477z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.844079 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5dbefe8-13da-4839-a4b2-b8864bb698a2" (UID: "e5dbefe8-13da-4839-a4b2-b8864bb698a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.883281 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-477z5\" (UniqueName: \"kubernetes.io/projected/e5dbefe8-13da-4839-a4b2-b8864bb698a2-kube-api-access-477z5\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.883607 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:07 crc kubenswrapper[5002]: I1014 08:15:07.883710 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5dbefe8-13da-4839-a4b2-b8864bb698a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.196394 5002 generic.go:334] "Generic (PLEG): container finished" podID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerID="4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671" exitCode=0 Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.196445 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerDied","Data":"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671"} Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.196461 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwsdx" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.196485 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwsdx" event={"ID":"e5dbefe8-13da-4839-a4b2-b8864bb698a2","Type":"ContainerDied","Data":"0accd0436d78e890cb81ecb8128d140fee883dfbf6fbf9ea4a89661f26960be7"} Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.196504 5002 scope.go:117] "RemoveContainer" containerID="4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.224816 5002 scope.go:117] "RemoveContainer" containerID="d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.245544 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.253635 5002 scope.go:117] "RemoveContainer" containerID="ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.259789 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lwsdx"] Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.293670 5002 scope.go:117] "RemoveContainer" containerID="4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671" Oct 14 08:15:08 crc kubenswrapper[5002]: E1014 08:15:08.294443 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671\": container with ID starting with 4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671 not found: ID does not exist" containerID="4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.294520 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671"} err="failed to get container status \"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671\": rpc error: code = NotFound desc = could not find container \"4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671\": container with ID starting with 4ba51e02561fe2c4c521d3ee7798f475cb10029532fdb6acd4c66785f0904671 not found: ID does not exist" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.294573 5002 scope.go:117] "RemoveContainer" containerID="d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1" Oct 14 08:15:08 crc kubenswrapper[5002]: E1014 08:15:08.295553 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1\": container with ID starting with d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1 not found: ID does not exist" containerID="d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.295588 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1"} err="failed to get container status \"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1\": rpc error: code = NotFound desc = could not find container \"d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1\": container with ID starting with d932ef8eec5617529d2d0add4ebe041bee5184482b8eedc55e8fff9ae9cbaeb1 not found: ID does not exist" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.295615 5002 scope.go:117] "RemoveContainer" containerID="ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3" Oct 14 08:15:08 crc kubenswrapper[5002]: E1014 08:15:08.295895 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3\": container with ID starting with ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3 not found: ID does not exist" containerID="ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3" Oct 14 08:15:08 crc kubenswrapper[5002]: I1014 08:15:08.295936 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3"} err="failed to get container status \"ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3\": rpc error: code = NotFound desc = could not find container \"ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3\": container with ID starting with ddb22a07e064cf6cc5e66d7316277785148808eb696163fdcf4de52206ebbae3 not found: ID does not exist" Oct 14 08:15:09 crc kubenswrapper[5002]: I1014 08:15:09.218338 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:15:09 crc kubenswrapper[5002]: I1014 08:15:09.218772 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:15:09 crc kubenswrapper[5002]: I1014 08:15:09.741101 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" path="/var/lib/kubelet/pods/e5dbefe8-13da-4839-a4b2-b8864bb698a2/volumes" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.869090 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:19 crc kubenswrapper[5002]: E1014 08:15:19.870269 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efd0fd5-8a8f-401f-a14f-ea0dee615274" containerName="collect-profiles" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870286 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efd0fd5-8a8f-401f-a14f-ea0dee615274" containerName="collect-profiles" Oct 14 08:15:19 crc kubenswrapper[5002]: E1014 08:15:19.870298 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="registry-server" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870304 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="registry-server" Oct 14 08:15:19 crc kubenswrapper[5002]: E1014 08:15:19.870319 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="extract-content" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870325 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="extract-content" Oct 14 08:15:19 crc kubenswrapper[5002]: E1014 08:15:19.870340 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="extract-utilities" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870346 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="extract-utilities" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870537 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efd0fd5-8a8f-401f-a14f-ea0dee615274" containerName="collect-profiles" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.870553 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5dbefe8-13da-4839-a4b2-b8864bb698a2" containerName="registry-server" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.871797 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.892559 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.947012 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.947110 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:19 crc kubenswrapper[5002]: I1014 08:15:19.947282 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfpjs\" (UniqueName: \"kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.049435 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.049507 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.049553 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfpjs\" (UniqueName: \"kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.049982 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.050057 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.066948 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfpjs\" (UniqueName: \"kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs\") pod \"community-operators-bttk4\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.188042 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:20 crc kubenswrapper[5002]: I1014 08:15:20.704796 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:21 crc kubenswrapper[5002]: I1014 08:15:21.362017 5002 generic.go:334] "Generic (PLEG): container finished" podID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerID="a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a" exitCode=0 Oct 14 08:15:21 crc kubenswrapper[5002]: I1014 08:15:21.362129 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerDied","Data":"a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a"} Oct 14 08:15:21 crc kubenswrapper[5002]: I1014 08:15:21.362346 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerStarted","Data":"b9fd8915c9733eaf0f7b729539fc2831053c9d3e7d090e25dc5b60282c2bb54c"} Oct 14 08:15:22 crc kubenswrapper[5002]: I1014 08:15:22.371711 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerStarted","Data":"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7"} Oct 14 08:15:23 crc kubenswrapper[5002]: I1014 08:15:23.389072 5002 generic.go:334] "Generic (PLEG): container finished" podID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerID="9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7" exitCode=0 Oct 14 08:15:23 crc kubenswrapper[5002]: I1014 08:15:23.389178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerDied","Data":"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7"} Oct 14 08:15:24 crc kubenswrapper[5002]: I1014 08:15:24.400501 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerStarted","Data":"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566"} Oct 14 08:15:24 crc kubenswrapper[5002]: I1014 08:15:24.421762 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bttk4" podStartSLOduration=2.937023618 podStartE2EDuration="5.421745535s" podCreationTimestamp="2025-10-14 08:15:19 +0000 UTC" firstStartedPulling="2025-10-14 08:15:21.364412778 +0000 UTC m=+1454.345652240" lastFinishedPulling="2025-10-14 08:15:23.849134665 +0000 UTC m=+1456.830374157" observedRunningTime="2025-10-14 08:15:24.418482107 +0000 UTC m=+1457.399721569" watchObservedRunningTime="2025-10-14 08:15:24.421745535 +0000 UTC m=+1457.402984977" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.257976 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.261284 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.274469 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.314117 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.314247 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.314421 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65xxn\" (UniqueName: \"kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.424184 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.424250 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.424318 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65xxn\" (UniqueName: \"kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.425094 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.425349 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.457217 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65xxn\" (UniqueName: \"kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn\") pod \"redhat-operators-scjg8\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:28 crc kubenswrapper[5002]: I1014 08:15:28.596271 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.092990 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:29 crc kubenswrapper[5002]: W1014 08:15:29.099021 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce2ef1b3_960d_4d82_8d6f_496fb4c9b4ac.slice/crio-51c540a7a679f69be7e425504975219e38458e1ff79c1758d6b1c7b7f6d13fb5 WatchSource:0}: Error finding container 51c540a7a679f69be7e425504975219e38458e1ff79c1758d6b1c7b7f6d13fb5: Status 404 returned error can't find the container with id 51c540a7a679f69be7e425504975219e38458e1ff79c1758d6b1c7b7f6d13fb5 Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.459537 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerID="7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630" exitCode=0 Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.459609 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerDied","Data":"7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630"} Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.459686 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerStarted","Data":"51c540a7a679f69be7e425504975219e38458e1ff79c1758d6b1c7b7f6d13fb5"} Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.462561 5002 generic.go:334] "Generic (PLEG): container finished" podID="e5f087b0-c7ff-46d2-8134-94f0879ff23c" containerID="3371c2d98c7c31e767da693319499d2d3528e35a8c95219bf4615f157bdd9c81" exitCode=0 Oct 14 08:15:29 crc kubenswrapper[5002]: I1014 08:15:29.462601 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" event={"ID":"e5f087b0-c7ff-46d2-8134-94f0879ff23c","Type":"ContainerDied","Data":"3371c2d98c7c31e767da693319499d2d3528e35a8c95219bf4615f157bdd9c81"} Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.188468 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.189199 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.266833 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.477160 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerStarted","Data":"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba"} Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.590117 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.934509 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.974549 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle\") pod \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.974659 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key\") pod \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.974897 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory\") pod \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.974946 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wdnn\" (UniqueName: \"kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn\") pod \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\" (UID: \"e5f087b0-c7ff-46d2-8134-94f0879ff23c\") " Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.984398 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e5f087b0-c7ff-46d2-8134-94f0879ff23c" (UID: "e5f087b0-c7ff-46d2-8134-94f0879ff23c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:30 crc kubenswrapper[5002]: I1014 08:15:30.985011 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn" (OuterVolumeSpecName: "kube-api-access-9wdnn") pod "e5f087b0-c7ff-46d2-8134-94f0879ff23c" (UID: "e5f087b0-c7ff-46d2-8134-94f0879ff23c"). InnerVolumeSpecName "kube-api-access-9wdnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.012020 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory" (OuterVolumeSpecName: "inventory") pod "e5f087b0-c7ff-46d2-8134-94f0879ff23c" (UID: "e5f087b0-c7ff-46d2-8134-94f0879ff23c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.013626 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5f087b0-c7ff-46d2-8134-94f0879ff23c" (UID: "e5f087b0-c7ff-46d2-8134-94f0879ff23c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.077242 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.077464 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wdnn\" (UniqueName: \"kubernetes.io/projected/e5f087b0-c7ff-46d2-8134-94f0879ff23c-kube-api-access-9wdnn\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.077542 5002 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.077618 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f087b0-c7ff-46d2-8134-94f0879ff23c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.495606 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" event={"ID":"e5f087b0-c7ff-46d2-8134-94f0879ff23c","Type":"ContainerDied","Data":"7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509"} Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.497266 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7720e3ae668d161968e04ff97d264331c3389f26331002aa431ac05179f7b509" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.495711 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.602393 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq"] Oct 14 08:15:31 crc kubenswrapper[5002]: E1014 08:15:31.602913 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f087b0-c7ff-46d2-8134-94f0879ff23c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.602948 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f087b0-c7ff-46d2-8134-94f0879ff23c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.603258 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f087b0-c7ff-46d2-8134-94f0879ff23c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.604160 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.606176 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.606645 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.607795 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.609012 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.629307 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq"] Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.687816 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.687908 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.687988 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slc8w\" (UniqueName: \"kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.789160 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slc8w\" (UniqueName: \"kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.789314 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.789507 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.794739 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.795452 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.811488 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slc8w\" (UniqueName: \"kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:31 crc kubenswrapper[5002]: I1014 08:15:31.935546 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:15:32 crc kubenswrapper[5002]: I1014 08:15:32.503739 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq"] Oct 14 08:15:32 crc kubenswrapper[5002]: I1014 08:15:32.513273 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerID="e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba" exitCode=0 Oct 14 08:15:32 crc kubenswrapper[5002]: I1014 08:15:32.513315 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerDied","Data":"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba"} Oct 14 08:15:32 crc kubenswrapper[5002]: I1014 08:15:32.640102 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:33 crc kubenswrapper[5002]: I1014 08:15:33.526574 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" event={"ID":"929d05f8-81d3-485e-a16b-50c701235a7a","Type":"ContainerStarted","Data":"3b1d2eed4a1fed5ae355ab406feaa2b569a9113bffb7f0f5b0fe08713613ecf8"} Oct 14 08:15:33 crc kubenswrapper[5002]: I1014 08:15:33.526738 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bttk4" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="registry-server" containerID="cri-o://ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566" gracePeriod=2 Oct 14 08:15:33 crc kubenswrapper[5002]: I1014 08:15:33.950618 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.032064 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfpjs\" (UniqueName: \"kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs\") pod \"5876aba6-c0b3-490c-bdf4-d83ec6592570\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.032384 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities\") pod \"5876aba6-c0b3-490c-bdf4-d83ec6592570\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.033136 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities" (OuterVolumeSpecName: "utilities") pod "5876aba6-c0b3-490c-bdf4-d83ec6592570" (UID: "5876aba6-c0b3-490c-bdf4-d83ec6592570"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.033191 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content\") pod \"5876aba6-c0b3-490c-bdf4-d83ec6592570\" (UID: \"5876aba6-c0b3-490c-bdf4-d83ec6592570\") " Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.037614 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs" (OuterVolumeSpecName: "kube-api-access-jfpjs") pod "5876aba6-c0b3-490c-bdf4-d83ec6592570" (UID: "5876aba6-c0b3-490c-bdf4-d83ec6592570"). InnerVolumeSpecName "kube-api-access-jfpjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.045817 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.045896 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfpjs\" (UniqueName: \"kubernetes.io/projected/5876aba6-c0b3-490c-bdf4-d83ec6592570-kube-api-access-jfpjs\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.085627 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5876aba6-c0b3-490c-bdf4-d83ec6592570" (UID: "5876aba6-c0b3-490c-bdf4-d83ec6592570"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.147547 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5876aba6-c0b3-490c-bdf4-d83ec6592570-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.537817 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerStarted","Data":"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f"} Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.540044 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" event={"ID":"929d05f8-81d3-485e-a16b-50c701235a7a","Type":"ContainerStarted","Data":"ed220500ec6be424d67c9d0da1600496d23405619f8f1757e321531c68117089"} Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.544607 5002 generic.go:334] "Generic (PLEG): container finished" podID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerID="ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566" exitCode=0 Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.544660 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bttk4" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.544672 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerDied","Data":"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566"} Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.544718 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bttk4" event={"ID":"5876aba6-c0b3-490c-bdf4-d83ec6592570","Type":"ContainerDied","Data":"b9fd8915c9733eaf0f7b729539fc2831053c9d3e7d090e25dc5b60282c2bb54c"} Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.544761 5002 scope.go:117] "RemoveContainer" containerID="ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.595241 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-scjg8" podStartSLOduration=2.726672546 podStartE2EDuration="6.595214565s" podCreationTimestamp="2025-10-14 08:15:28 +0000 UTC" firstStartedPulling="2025-10-14 08:15:29.462185189 +0000 UTC m=+1462.443424641" lastFinishedPulling="2025-10-14 08:15:33.330727198 +0000 UTC m=+1466.311966660" observedRunningTime="2025-10-14 08:15:34.577526317 +0000 UTC m=+1467.558765799" watchObservedRunningTime="2025-10-14 08:15:34.595214565 +0000 UTC m=+1467.576454067" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.597523 5002 scope.go:117] "RemoveContainer" containerID="9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.611940 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.617829 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bttk4"] Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.622387 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" podStartSLOduration=2.615977011 podStartE2EDuration="3.62237388s" podCreationTimestamp="2025-10-14 08:15:31 +0000 UTC" firstStartedPulling="2025-10-14 08:15:32.521141311 +0000 UTC m=+1465.502380803" lastFinishedPulling="2025-10-14 08:15:33.52753822 +0000 UTC m=+1466.508777672" observedRunningTime="2025-10-14 08:15:34.614102056 +0000 UTC m=+1467.595341508" watchObservedRunningTime="2025-10-14 08:15:34.62237388 +0000 UTC m=+1467.603613322" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.646358 5002 scope.go:117] "RemoveContainer" containerID="a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.670880 5002 scope.go:117] "RemoveContainer" containerID="ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566" Oct 14 08:15:34 crc kubenswrapper[5002]: E1014 08:15:34.678282 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566\": container with ID starting with ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566 not found: ID does not exist" containerID="ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.678337 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566"} err="failed to get container status \"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566\": rpc error: code = NotFound desc = could not find container \"ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566\": container with ID starting with ac68c0091a37446db61945c7ad31ee8ce4c89242f270444b0ec199d03d11a566 not found: ID does not exist" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.678373 5002 scope.go:117] "RemoveContainer" containerID="9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7" Oct 14 08:15:34 crc kubenswrapper[5002]: E1014 08:15:34.678949 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7\": container with ID starting with 9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7 not found: ID does not exist" containerID="9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.678993 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7"} err="failed to get container status \"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7\": rpc error: code = NotFound desc = could not find container \"9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7\": container with ID starting with 9e391b17bd5615624c842d616fd73cabf1dd5ead2b83720e35ba8d5ed64f74b7 not found: ID does not exist" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.679019 5002 scope.go:117] "RemoveContainer" containerID="a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a" Oct 14 08:15:34 crc kubenswrapper[5002]: E1014 08:15:34.679267 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a\": container with ID starting with a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a not found: ID does not exist" containerID="a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a" Oct 14 08:15:34 crc kubenswrapper[5002]: I1014 08:15:34.679307 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a"} err="failed to get container status \"a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a\": rpc error: code = NotFound desc = could not find container \"a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a\": container with ID starting with a348b717d096e2e415798ada7bf7a6c1a0c31bbf942ccdcf6fcafc90f712a11a not found: ID does not exist" Oct 14 08:15:35 crc kubenswrapper[5002]: I1014 08:15:35.730382 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" path="/var/lib/kubelet/pods/5876aba6-c0b3-490c-bdf4-d83ec6592570/volumes" Oct 14 08:15:38 crc kubenswrapper[5002]: I1014 08:15:38.596536 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:38 crc kubenswrapper[5002]: I1014 08:15:38.596789 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:38 crc kubenswrapper[5002]: I1014 08:15:38.651958 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.218140 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.218485 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.218552 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.219657 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.219876 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" gracePeriod=600 Oct 14 08:15:39 crc kubenswrapper[5002]: E1014 08:15:39.352546 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.591353 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" exitCode=0 Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.591437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63"} Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.591486 5002 scope.go:117] "RemoveContainer" containerID="07abb25201954755af814e2503a7a4a5d72078b65a4908d4f0261eac6ce52cf5" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.591937 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:15:39 crc kubenswrapper[5002]: E1014 08:15:39.592236 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.676582 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:39 crc kubenswrapper[5002]: I1014 08:15:39.740178 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:41 crc kubenswrapper[5002]: I1014 08:15:41.615808 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-scjg8" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="registry-server" containerID="cri-o://b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f" gracePeriod=2 Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.155888 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.222399 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content\") pod \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.222485 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65xxn\" (UniqueName: \"kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn\") pod \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.222582 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities\") pod \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\" (UID: \"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac\") " Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.223640 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities" (OuterVolumeSpecName: "utilities") pod "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" (UID: "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.231430 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn" (OuterVolumeSpecName: "kube-api-access-65xxn") pod "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" (UID: "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac"). InnerVolumeSpecName "kube-api-access-65xxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.324397 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65xxn\" (UniqueName: \"kubernetes.io/projected/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-kube-api-access-65xxn\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.324428 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.333252 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" (UID: "ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.426878 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.637718 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerID="b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f" exitCode=0 Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.637771 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerDied","Data":"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f"} Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.637801 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-scjg8" event={"ID":"ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac","Type":"ContainerDied","Data":"51c540a7a679f69be7e425504975219e38458e1ff79c1758d6b1c7b7f6d13fb5"} Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.637811 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-scjg8" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.637817 5002 scope.go:117] "RemoveContainer" containerID="b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.678028 5002 scope.go:117] "RemoveContainer" containerID="e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.718535 5002 scope.go:117] "RemoveContainer" containerID="7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.721560 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.738417 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-scjg8"] Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.763747 5002 scope.go:117] "RemoveContainer" containerID="b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f" Oct 14 08:15:42 crc kubenswrapper[5002]: E1014 08:15:42.764304 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f\": container with ID starting with b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f not found: ID does not exist" containerID="b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.764334 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f"} err="failed to get container status \"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f\": rpc error: code = NotFound desc = could not find container \"b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f\": container with ID starting with b8177a24ec2c667968b49acba5d6f6dc1fac27065b1732ff1f312824b859f71f not found: ID does not exist" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.764356 5002 scope.go:117] "RemoveContainer" containerID="e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba" Oct 14 08:15:42 crc kubenswrapper[5002]: E1014 08:15:42.764581 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba\": container with ID starting with e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba not found: ID does not exist" containerID="e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.764607 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba"} err="failed to get container status \"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba\": rpc error: code = NotFound desc = could not find container \"e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba\": container with ID starting with e29b8f80a675209dc6147e0cd7a6629deb114a9ef527894dce8835f4f6aa35ba not found: ID does not exist" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.764621 5002 scope.go:117] "RemoveContainer" containerID="7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630" Oct 14 08:15:42 crc kubenswrapper[5002]: E1014 08:15:42.764812 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630\": container with ID starting with 7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630 not found: ID does not exist" containerID="7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630" Oct 14 08:15:42 crc kubenswrapper[5002]: I1014 08:15:42.764849 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630"} err="failed to get container status \"7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630\": rpc error: code = NotFound desc = could not find container \"7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630\": container with ID starting with 7cc72001dc0b65fafb870e9be8981125703bace346d03dbad11641698f481630 not found: ID does not exist" Oct 14 08:15:43 crc kubenswrapper[5002]: I1014 08:15:43.729572 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" path="/var/lib/kubelet/pods/ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac/volumes" Oct 14 08:15:52 crc kubenswrapper[5002]: I1014 08:15:52.721439 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:15:52 crc kubenswrapper[5002]: E1014 08:15:52.722741 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:07 crc kubenswrapper[5002]: I1014 08:16:07.736481 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:16:07 crc kubenswrapper[5002]: E1014 08:16:07.737662 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:20 crc kubenswrapper[5002]: I1014 08:16:20.720644 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:16:20 crc kubenswrapper[5002]: E1014 08:16:20.721827 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:22 crc kubenswrapper[5002]: I1014 08:16:22.126577 5002 scope.go:117] "RemoveContainer" containerID="1f01369e35849a39a781ed5d97212578aa7f7446321007e455568f44d3056513" Oct 14 08:16:31 crc kubenswrapper[5002]: I1014 08:16:31.720870 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:16:31 crc kubenswrapper[5002]: E1014 08:16:31.721804 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:46 crc kubenswrapper[5002]: I1014 08:16:46.098671 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vgqc8"] Oct 14 08:16:46 crc kubenswrapper[5002]: I1014 08:16:46.107994 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-md8d9"] Oct 14 08:16:46 crc kubenswrapper[5002]: I1014 08:16:46.116493 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vgqc8"] Oct 14 08:16:46 crc kubenswrapper[5002]: I1014 08:16:46.126202 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-md8d9"] Oct 14 08:16:46 crc kubenswrapper[5002]: I1014 08:16:46.720810 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:16:46 crc kubenswrapper[5002]: E1014 08:16:46.721700 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:47 crc kubenswrapper[5002]: I1014 08:16:47.734553 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb5f6d1-6253-4c86-9953-b83e40c97276" path="/var/lib/kubelet/pods/6eb5f6d1-6253-4c86-9953-b83e40c97276/volumes" Oct 14 08:16:47 crc kubenswrapper[5002]: I1014 08:16:47.735682 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70b9e9fc-6eb1-4948-9f50-c9245db17ea8" path="/var/lib/kubelet/pods/70b9e9fc-6eb1-4948-9f50-c9245db17ea8/volumes" Oct 14 08:16:49 crc kubenswrapper[5002]: I1014 08:16:49.437432 5002 generic.go:334] "Generic (PLEG): container finished" podID="929d05f8-81d3-485e-a16b-50c701235a7a" containerID="ed220500ec6be424d67c9d0da1600496d23405619f8f1757e321531c68117089" exitCode=0 Oct 14 08:16:49 crc kubenswrapper[5002]: I1014 08:16:49.437860 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" event={"ID":"929d05f8-81d3-485e-a16b-50c701235a7a","Type":"ContainerDied","Data":"ed220500ec6be424d67c9d0da1600496d23405619f8f1757e321531c68117089"} Oct 14 08:16:50 crc kubenswrapper[5002]: I1014 08:16:50.971960 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.083993 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slc8w\" (UniqueName: \"kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w\") pod \"929d05f8-81d3-485e-a16b-50c701235a7a\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.084116 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory\") pod \"929d05f8-81d3-485e-a16b-50c701235a7a\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.084203 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key\") pod \"929d05f8-81d3-485e-a16b-50c701235a7a\" (UID: \"929d05f8-81d3-485e-a16b-50c701235a7a\") " Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.092624 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w" (OuterVolumeSpecName: "kube-api-access-slc8w") pod "929d05f8-81d3-485e-a16b-50c701235a7a" (UID: "929d05f8-81d3-485e-a16b-50c701235a7a"). InnerVolumeSpecName "kube-api-access-slc8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.121824 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "929d05f8-81d3-485e-a16b-50c701235a7a" (UID: "929d05f8-81d3-485e-a16b-50c701235a7a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.138547 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory" (OuterVolumeSpecName: "inventory") pod "929d05f8-81d3-485e-a16b-50c701235a7a" (UID: "929d05f8-81d3-485e-a16b-50c701235a7a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.186780 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slc8w\" (UniqueName: \"kubernetes.io/projected/929d05f8-81d3-485e-a16b-50c701235a7a-kube-api-access-slc8w\") on node \"crc\" DevicePath \"\"" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.186816 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.186825 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/929d05f8-81d3-485e-a16b-50c701235a7a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.465545 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" event={"ID":"929d05f8-81d3-485e-a16b-50c701235a7a","Type":"ContainerDied","Data":"3b1d2eed4a1fed5ae355ab406feaa2b569a9113bffb7f0f5b0fe08713613ecf8"} Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.465596 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1d2eed4a1fed5ae355ab406feaa2b569a9113bffb7f0f5b0fe08713613ecf8" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.465661 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.566319 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869"] Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.566925 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="extract-utilities" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.566966 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="extract-utilities" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567003 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="extract-content" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567012 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="extract-content" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567034 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567042 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567059 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567095 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567110 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="929d05f8-81d3-485e-a16b-50c701235a7a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567120 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="929d05f8-81d3-485e-a16b-50c701235a7a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567138 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="extract-content" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567145 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="extract-content" Oct 14 08:16:51 crc kubenswrapper[5002]: E1014 08:16:51.567165 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="extract-utilities" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.567173 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="extract-utilities" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.571298 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="929d05f8-81d3-485e-a16b-50c701235a7a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.571353 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2ef1b3-960d-4d82-8d6f-496fb4c9b4ac" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.571372 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="5876aba6-c0b3-490c-bdf4-d83ec6592570" containerName="registry-server" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.572030 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.574109 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.574479 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.574681 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.575934 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.579698 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869"] Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.704201 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.704316 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.704408 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h2fs\" (UniqueName: \"kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.806349 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h2fs\" (UniqueName: \"kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.806523 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.806723 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.810795 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.811756 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.837250 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h2fs\" (UniqueName: \"kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-kg869\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:51 crc kubenswrapper[5002]: I1014 08:16:51.953143 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:16:52 crc kubenswrapper[5002]: I1014 08:16:52.037587 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rhqxc"] Oct 14 08:16:52 crc kubenswrapper[5002]: I1014 08:16:52.055757 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rhqxc"] Oct 14 08:16:52 crc kubenswrapper[5002]: I1014 08:16:52.479256 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869"] Oct 14 08:16:52 crc kubenswrapper[5002]: W1014 08:16:52.486576 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda88843ae_ab49_49f1_bc52_f82a2ad016ec.slice/crio-bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424 WatchSource:0}: Error finding container bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424: Status 404 returned error can't find the container with id bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424 Oct 14 08:16:53 crc kubenswrapper[5002]: I1014 08:16:53.485637 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" event={"ID":"a88843ae-ab49-49f1-bc52-f82a2ad016ec","Type":"ContainerStarted","Data":"0af358cc8b22a0d13a188104cee46c4e8c77e58bdc3862f7bacaeee066260fc3"} Oct 14 08:16:53 crc kubenswrapper[5002]: I1014 08:16:53.486125 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" event={"ID":"a88843ae-ab49-49f1-bc52-f82a2ad016ec","Type":"ContainerStarted","Data":"bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424"} Oct 14 08:16:53 crc kubenswrapper[5002]: I1014 08:16:53.511674 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" podStartSLOduration=2.059920436 podStartE2EDuration="2.51165609s" podCreationTimestamp="2025-10-14 08:16:51 +0000 UTC" firstStartedPulling="2025-10-14 08:16:52.48983018 +0000 UTC m=+1545.471069662" lastFinishedPulling="2025-10-14 08:16:52.941565824 +0000 UTC m=+1545.922805316" observedRunningTime="2025-10-14 08:16:53.504061825 +0000 UTC m=+1546.485301287" watchObservedRunningTime="2025-10-14 08:16:53.51165609 +0000 UTC m=+1546.492895552" Oct 14 08:16:53 crc kubenswrapper[5002]: I1014 08:16:53.738478 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb1ba9e-3684-4c6a-a33a-82e316a44cd6" path="/var/lib/kubelet/pods/cfb1ba9e-3684-4c6a-a33a-82e316a44cd6/volumes" Oct 14 08:16:56 crc kubenswrapper[5002]: I1014 08:16:56.041830 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-44d2-account-create-2k92k"] Oct 14 08:16:56 crc kubenswrapper[5002]: I1014 08:16:56.060671 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-a057-account-create-rgqxv"] Oct 14 08:16:56 crc kubenswrapper[5002]: I1014 08:16:56.068661 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-44d2-account-create-2k92k"] Oct 14 08:16:56 crc kubenswrapper[5002]: I1014 08:16:56.075203 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-a057-account-create-rgqxv"] Oct 14 08:16:57 crc kubenswrapper[5002]: I1014 08:16:57.732737 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:16:57 crc kubenswrapper[5002]: I1014 08:16:57.732951 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4e8168d-4cbe-4032-9ab8-02a8c072a5b9" path="/var/lib/kubelet/pods/a4e8168d-4cbe-4032-9ab8-02a8c072a5b9/volumes" Oct 14 08:16:57 crc kubenswrapper[5002]: E1014 08:16:57.733609 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:16:57 crc kubenswrapper[5002]: I1014 08:16:57.733806 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc62cd95-ad2f-4f14-820f-8545fb48080b" path="/var/lib/kubelet/pods/cc62cd95-ad2f-4f14-820f-8545fb48080b/volumes" Oct 14 08:16:58 crc kubenswrapper[5002]: I1014 08:16:58.566919 5002 generic.go:334] "Generic (PLEG): container finished" podID="a88843ae-ab49-49f1-bc52-f82a2ad016ec" containerID="0af358cc8b22a0d13a188104cee46c4e8c77e58bdc3862f7bacaeee066260fc3" exitCode=0 Oct 14 08:16:58 crc kubenswrapper[5002]: I1014 08:16:58.566989 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" event={"ID":"a88843ae-ab49-49f1-bc52-f82a2ad016ec","Type":"ContainerDied","Data":"0af358cc8b22a0d13a188104cee46c4e8c77e58bdc3862f7bacaeee066260fc3"} Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.131922 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.291265 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h2fs\" (UniqueName: \"kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs\") pod \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.291342 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key\") pod \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.292683 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory\") pod \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\" (UID: \"a88843ae-ab49-49f1-bc52-f82a2ad016ec\") " Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.301670 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs" (OuterVolumeSpecName: "kube-api-access-6h2fs") pod "a88843ae-ab49-49f1-bc52-f82a2ad016ec" (UID: "a88843ae-ab49-49f1-bc52-f82a2ad016ec"). InnerVolumeSpecName "kube-api-access-6h2fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.329918 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a88843ae-ab49-49f1-bc52-f82a2ad016ec" (UID: "a88843ae-ab49-49f1-bc52-f82a2ad016ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.337767 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory" (OuterVolumeSpecName: "inventory") pod "a88843ae-ab49-49f1-bc52-f82a2ad016ec" (UID: "a88843ae-ab49-49f1-bc52-f82a2ad016ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.395357 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h2fs\" (UniqueName: \"kubernetes.io/projected/a88843ae-ab49-49f1-bc52-f82a2ad016ec-kube-api-access-6h2fs\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.395413 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.395429 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a88843ae-ab49-49f1-bc52-f82a2ad016ec-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.594260 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" event={"ID":"a88843ae-ab49-49f1-bc52-f82a2ad016ec","Type":"ContainerDied","Data":"bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424"} Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.594307 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfd0f61a4b29ab80d153a64724e85d3a5b1a9ff8587109bb89b8d2e73eb7b424" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.594956 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.694712 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p"] Oct 14 08:17:00 crc kubenswrapper[5002]: E1014 08:17:00.695189 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88843ae-ab49-49f1-bc52-f82a2ad016ec" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.695211 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88843ae-ab49-49f1-bc52-f82a2ad016ec" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.695457 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88843ae-ab49-49f1-bc52-f82a2ad016ec" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.696235 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.702412 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.704927 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.705038 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.705388 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.711973 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p"] Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.803378 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr8df\" (UniqueName: \"kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.803464 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.803561 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.906271 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr8df\" (UniqueName: \"kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.906691 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.907376 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.916226 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.919996 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:00 crc kubenswrapper[5002]: I1014 08:17:00.927954 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr8df\" (UniqueName: \"kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-m2l2p\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:01 crc kubenswrapper[5002]: I1014 08:17:01.029113 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:01 crc kubenswrapper[5002]: I1014 08:17:01.669553 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p"] Oct 14 08:17:01 crc kubenswrapper[5002]: I1014 08:17:01.676678 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:17:02 crc kubenswrapper[5002]: I1014 08:17:02.041174 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f8d8-account-create-f78dk"] Oct 14 08:17:02 crc kubenswrapper[5002]: I1014 08:17:02.080813 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f8d8-account-create-f78dk"] Oct 14 08:17:02 crc kubenswrapper[5002]: I1014 08:17:02.626560 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" event={"ID":"3c014d84-bdb3-4e4a-b47f-00e315ed4023","Type":"ContainerStarted","Data":"20850f97df7ef558d1ba5adfc9b22d26cf093d495225451dcce1629b566540f5"} Oct 14 08:17:02 crc kubenswrapper[5002]: I1014 08:17:02.626867 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" event={"ID":"3c014d84-bdb3-4e4a-b47f-00e315ed4023","Type":"ContainerStarted","Data":"ec198b4d0a8e78b3caa258d4d2b119e460a74a5d4aff3b14bdb1e01053bedbf5"} Oct 14 08:17:02 crc kubenswrapper[5002]: I1014 08:17:02.652559 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" podStartSLOduration=2.197665616 podStartE2EDuration="2.652530536s" podCreationTimestamp="2025-10-14 08:17:00 +0000 UTC" firstStartedPulling="2025-10-14 08:17:01.676432781 +0000 UTC m=+1554.657672233" lastFinishedPulling="2025-10-14 08:17:02.131297691 +0000 UTC m=+1555.112537153" observedRunningTime="2025-10-14 08:17:02.64716696 +0000 UTC m=+1555.628406442" watchObservedRunningTime="2025-10-14 08:17:02.652530536 +0000 UTC m=+1555.633770028" Oct 14 08:17:03 crc kubenswrapper[5002]: I1014 08:17:03.741119 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed027b3-8a30-4fee-84c0-70e03887479a" path="/var/lib/kubelet/pods/7ed027b3-8a30-4fee-84c0-70e03887479a/volumes" Oct 14 08:17:11 crc kubenswrapper[5002]: I1014 08:17:11.721430 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:17:11 crc kubenswrapper[5002]: E1014 08:17:11.722372 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:17:18 crc kubenswrapper[5002]: I1014 08:17:18.073927 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-snvl8"] Oct 14 08:17:18 crc kubenswrapper[5002]: I1014 08:17:18.092053 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-l8s85"] Oct 14 08:17:18 crc kubenswrapper[5002]: I1014 08:17:18.102593 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-snvl8"] Oct 14 08:17:18 crc kubenswrapper[5002]: I1014 08:17:18.111483 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-l8s85"] Oct 14 08:17:19 crc kubenswrapper[5002]: I1014 08:17:19.735764 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f298eb8-cf82-4d09-93a2-f70fd6e50b9b" path="/var/lib/kubelet/pods/9f298eb8-cf82-4d09-93a2-f70fd6e50b9b/volumes" Oct 14 08:17:19 crc kubenswrapper[5002]: I1014 08:17:19.736541 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3cb8568-04da-484f-a4f8-f1b2e8ab755c" path="/var/lib/kubelet/pods/a3cb8568-04da-484f-a4f8-f1b2e8ab755c/volumes" Oct 14 08:17:21 crc kubenswrapper[5002]: I1014 08:17:21.048618 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-b87gc"] Oct 14 08:17:21 crc kubenswrapper[5002]: I1014 08:17:21.061422 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-b87gc"] Oct 14 08:17:21 crc kubenswrapper[5002]: I1014 08:17:21.740318 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e29da0d6-d597-4318-ba7f-da9c4d32e518" path="/var/lib/kubelet/pods/e29da0d6-d597-4318-ba7f-da9c4d32e518/volumes" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.228120 5002 scope.go:117] "RemoveContainer" containerID="cae03b9ef2c61b476773b1d165be3ff50a774aa17a2b9d3c4f92dae7e0f7002a" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.248447 5002 scope.go:117] "RemoveContainer" containerID="e253ff5b25f1f40c213293e16d64f22a7c4b29da53dd2f51fcb96a91b1ea18f7" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.314636 5002 scope.go:117] "RemoveContainer" containerID="205c0cb1f3dad50fced5285aae956e6710f40ee547061abb544105521a1938d8" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.363969 5002 scope.go:117] "RemoveContainer" containerID="9eb968117a909bce92185225c7db43136e1ea202e8aa533835cde343baf64a66" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.405423 5002 scope.go:117] "RemoveContainer" containerID="9ae5d2a808b5331b090ba5eefd901689444a0e09a5779f1e1a9ff29e4c63e2f1" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.447739 5002 scope.go:117] "RemoveContainer" containerID="8fffffe2126815d7948a951b571e0c52ce80cc61ddcbbb78171ac5fc8cbd8be2" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.493380 5002 scope.go:117] "RemoveContainer" containerID="a74e90ead29362576d8f1a8c92b4414432d8a5911887362030cc101440dd71a6" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.524036 5002 scope.go:117] "RemoveContainer" containerID="a8368296357610b5e1d057d3b522f52da842ecb976c1b288052d100593fd4fa3" Oct 14 08:17:22 crc kubenswrapper[5002]: I1014 08:17:22.562911 5002 scope.go:117] "RemoveContainer" containerID="b02b0c07dcd5d213197baaac31e1f4092b3042aaa519f3228584e603a71e01c6" Oct 14 08:17:24 crc kubenswrapper[5002]: I1014 08:17:24.722041 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:17:24 crc kubenswrapper[5002]: E1014 08:17:24.722885 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:17:25 crc kubenswrapper[5002]: I1014 08:17:25.035600 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-z4thp"] Oct 14 08:17:25 crc kubenswrapper[5002]: I1014 08:17:25.046930 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-z4thp"] Oct 14 08:17:25 crc kubenswrapper[5002]: I1014 08:17:25.738216 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b80ca7-5604-4c84-96f5-1279ecf2abe7" path="/var/lib/kubelet/pods/55b80ca7-5604-4c84-96f5-1279ecf2abe7/volumes" Oct 14 08:17:26 crc kubenswrapper[5002]: I1014 08:17:26.044894 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4tkcz"] Oct 14 08:17:26 crc kubenswrapper[5002]: I1014 08:17:26.053237 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4tkcz"] Oct 14 08:17:27 crc kubenswrapper[5002]: I1014 08:17:27.738917 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa02e7c2-896a-4750-968d-9a4543651928" path="/var/lib/kubelet/pods/aa02e7c2-896a-4750-968d-9a4543651928/volumes" Oct 14 08:17:36 crc kubenswrapper[5002]: I1014 08:17:36.721508 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:17:36 crc kubenswrapper[5002]: E1014 08:17:36.722544 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.068194 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-56ab-account-create-64pzr"] Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.084542 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-78b8-account-create-cx2pn"] Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.095010 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-78b8-account-create-cx2pn"] Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.111786 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-56ab-account-create-64pzr"] Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.127139 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f03f-account-create-6kv52"] Oct 14 08:17:44 crc kubenswrapper[5002]: I1014 08:17:44.134760 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f03f-account-create-6kv52"] Oct 14 08:17:45 crc kubenswrapper[5002]: I1014 08:17:45.125819 5002 generic.go:334] "Generic (PLEG): container finished" podID="3c014d84-bdb3-4e4a-b47f-00e315ed4023" containerID="20850f97df7ef558d1ba5adfc9b22d26cf093d495225451dcce1629b566540f5" exitCode=0 Oct 14 08:17:45 crc kubenswrapper[5002]: I1014 08:17:45.126067 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" event={"ID":"3c014d84-bdb3-4e4a-b47f-00e315ed4023","Type":"ContainerDied","Data":"20850f97df7ef558d1ba5adfc9b22d26cf093d495225451dcce1629b566540f5"} Oct 14 08:17:45 crc kubenswrapper[5002]: I1014 08:17:45.743641 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b0300de-4b38-4615-afd4-910e29e28127" path="/var/lib/kubelet/pods/3b0300de-4b38-4615-afd4-910e29e28127/volumes" Oct 14 08:17:45 crc kubenswrapper[5002]: I1014 08:17:45.744710 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d1d7d9-db0a-47a7-b8de-23b98786f876" path="/var/lib/kubelet/pods/84d1d7d9-db0a-47a7-b8de-23b98786f876/volumes" Oct 14 08:17:45 crc kubenswrapper[5002]: I1014 08:17:45.745771 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8d232a-e9c2-404f-a09d-4c8e609c9b6a" path="/var/lib/kubelet/pods/ca8d232a-e9c2-404f-a09d-4c8e609c9b6a/volumes" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.671695 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.704614 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr8df\" (UniqueName: \"kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df\") pod \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.704874 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key\") pod \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.705035 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory\") pod \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\" (UID: \"3c014d84-bdb3-4e4a-b47f-00e315ed4023\") " Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.712989 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df" (OuterVolumeSpecName: "kube-api-access-sr8df") pod "3c014d84-bdb3-4e4a-b47f-00e315ed4023" (UID: "3c014d84-bdb3-4e4a-b47f-00e315ed4023"). InnerVolumeSpecName "kube-api-access-sr8df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.741089 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c014d84-bdb3-4e4a-b47f-00e315ed4023" (UID: "3c014d84-bdb3-4e4a-b47f-00e315ed4023"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.766663 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory" (OuterVolumeSpecName: "inventory") pod "3c014d84-bdb3-4e4a-b47f-00e315ed4023" (UID: "3c014d84-bdb3-4e4a-b47f-00e315ed4023"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.807392 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.807437 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr8df\" (UniqueName: \"kubernetes.io/projected/3c014d84-bdb3-4e4a-b47f-00e315ed4023-kube-api-access-sr8df\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:46 crc kubenswrapper[5002]: I1014 08:17:46.807456 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c014d84-bdb3-4e4a-b47f-00e315ed4023-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.152964 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" event={"ID":"3c014d84-bdb3-4e4a-b47f-00e315ed4023","Type":"ContainerDied","Data":"ec198b4d0a8e78b3caa258d4d2b119e460a74a5d4aff3b14bdb1e01053bedbf5"} Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.153027 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec198b4d0a8e78b3caa258d4d2b119e460a74a5d4aff3b14bdb1e01053bedbf5" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.153080 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.275245 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9"] Oct 14 08:17:47 crc kubenswrapper[5002]: E1014 08:17:47.275671 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c014d84-bdb3-4e4a-b47f-00e315ed4023" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.275691 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c014d84-bdb3-4e4a-b47f-00e315ed4023" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.275942 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c014d84-bdb3-4e4a-b47f-00e315ed4023" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.276615 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.280903 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.281290 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.281624 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.281707 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.299820 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9"] Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.421927 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.421992 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pq5q\" (UniqueName: \"kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.422985 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.523922 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.523964 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pq5q\" (UniqueName: \"kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.524068 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.528725 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.529005 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.566056 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pq5q\" (UniqueName: \"kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.610473 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:47 crc kubenswrapper[5002]: I1014 08:17:47.990786 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9"] Oct 14 08:17:48 crc kubenswrapper[5002]: W1014 08:17:48.002304 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff80a75_e146_479b_b4b3_74adcc4491fc.slice/crio-03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a WatchSource:0}: Error finding container 03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a: Status 404 returned error can't find the container with id 03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a Oct 14 08:17:48 crc kubenswrapper[5002]: I1014 08:17:48.167233 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" event={"ID":"2ff80a75-e146-479b-b4b3-74adcc4491fc","Type":"ContainerStarted","Data":"03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a"} Oct 14 08:17:48 crc kubenswrapper[5002]: I1014 08:17:48.721042 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:17:48 crc kubenswrapper[5002]: E1014 08:17:48.721632 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:17:49 crc kubenswrapper[5002]: I1014 08:17:49.178123 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" event={"ID":"2ff80a75-e146-479b-b4b3-74adcc4491fc","Type":"ContainerStarted","Data":"39ccaaafac8f404b9ba7b864a496f9664eda9132ba4eae1ab74f219f99aef7c7"} Oct 14 08:17:49 crc kubenswrapper[5002]: I1014 08:17:49.203860 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" podStartSLOduration=1.581657838 podStartE2EDuration="2.203820462s" podCreationTimestamp="2025-10-14 08:17:47 +0000 UTC" firstStartedPulling="2025-10-14 08:17:48.004911493 +0000 UTC m=+1600.986150955" lastFinishedPulling="2025-10-14 08:17:48.627074117 +0000 UTC m=+1601.608313579" observedRunningTime="2025-10-14 08:17:49.195900539 +0000 UTC m=+1602.177140021" watchObservedRunningTime="2025-10-14 08:17:49.203820462 +0000 UTC m=+1602.185059924" Oct 14 08:17:52 crc kubenswrapper[5002]: I1014 08:17:52.047473 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-snwzv"] Oct 14 08:17:52 crc kubenswrapper[5002]: I1014 08:17:52.061971 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-snwzv"] Oct 14 08:17:53 crc kubenswrapper[5002]: I1014 08:17:53.737895 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c28ed22-e6c7-4a08-ba4f-eb8052296a5f" path="/var/lib/kubelet/pods/9c28ed22-e6c7-4a08-ba4f-eb8052296a5f/volumes" Oct 14 08:17:54 crc kubenswrapper[5002]: I1014 08:17:54.243397 5002 generic.go:334] "Generic (PLEG): container finished" podID="2ff80a75-e146-479b-b4b3-74adcc4491fc" containerID="39ccaaafac8f404b9ba7b864a496f9664eda9132ba4eae1ab74f219f99aef7c7" exitCode=0 Oct 14 08:17:54 crc kubenswrapper[5002]: I1014 08:17:54.243441 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" event={"ID":"2ff80a75-e146-479b-b4b3-74adcc4491fc","Type":"ContainerDied","Data":"39ccaaafac8f404b9ba7b864a496f9664eda9132ba4eae1ab74f219f99aef7c7"} Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.659862 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.802417 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key\") pod \"2ff80a75-e146-479b-b4b3-74adcc4491fc\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.802668 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory\") pod \"2ff80a75-e146-479b-b4b3-74adcc4491fc\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.802884 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pq5q\" (UniqueName: \"kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q\") pod \"2ff80a75-e146-479b-b4b3-74adcc4491fc\" (UID: \"2ff80a75-e146-479b-b4b3-74adcc4491fc\") " Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.811089 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q" (OuterVolumeSpecName: "kube-api-access-8pq5q") pod "2ff80a75-e146-479b-b4b3-74adcc4491fc" (UID: "2ff80a75-e146-479b-b4b3-74adcc4491fc"). InnerVolumeSpecName "kube-api-access-8pq5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.830065 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory" (OuterVolumeSpecName: "inventory") pod "2ff80a75-e146-479b-b4b3-74adcc4491fc" (UID: "2ff80a75-e146-479b-b4b3-74adcc4491fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.831351 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ff80a75-e146-479b-b4b3-74adcc4491fc" (UID: "2ff80a75-e146-479b-b4b3-74adcc4491fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.906343 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.906387 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff80a75-e146-479b-b4b3-74adcc4491fc-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:55 crc kubenswrapper[5002]: I1014 08:17:55.906404 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pq5q\" (UniqueName: \"kubernetes.io/projected/2ff80a75-e146-479b-b4b3-74adcc4491fc-kube-api-access-8pq5q\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.265789 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.265884 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9" event={"ID":"2ff80a75-e146-479b-b4b3-74adcc4491fc","Type":"ContainerDied","Data":"03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a"} Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.266329 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03d028134cf886d7bf2b0e9ecb69b683c13b64ff3a9297e9cba82fe2460f237a" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.416002 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb"] Oct 14 08:17:56 crc kubenswrapper[5002]: E1014 08:17:56.416509 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff80a75-e146-479b-b4b3-74adcc4491fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.416536 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff80a75-e146-479b-b4b3-74adcc4491fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.416774 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff80a75-e146-479b-b4b3-74adcc4491fc" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.417579 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.419824 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.420067 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.420213 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.420396 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.427602 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb"] Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.518387 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8jb\" (UniqueName: \"kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.518573 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.518683 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.621141 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8jb\" (UniqueName: \"kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.621225 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.621285 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.628211 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.628976 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.650518 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8jb\" (UniqueName: \"kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:56 crc kubenswrapper[5002]: I1014 08:17:56.735669 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:17:57 crc kubenswrapper[5002]: I1014 08:17:57.350263 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb"] Oct 14 08:17:57 crc kubenswrapper[5002]: W1014 08:17:57.352334 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fd571d7_b282_49c2_984d_b37b510d0d48.slice/crio-facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768 WatchSource:0}: Error finding container facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768: Status 404 returned error can't find the container with id facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768 Oct 14 08:17:58 crc kubenswrapper[5002]: I1014 08:17:58.300893 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" event={"ID":"2fd571d7-b282-49c2-984d-b37b510d0d48","Type":"ContainerStarted","Data":"898c56916e4d9e7bcdb6dbf835e5b43360e691116a7f2ce8289f0e514e58c622"} Oct 14 08:17:58 crc kubenswrapper[5002]: I1014 08:17:58.301292 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" event={"ID":"2fd571d7-b282-49c2-984d-b37b510d0d48","Type":"ContainerStarted","Data":"facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768"} Oct 14 08:17:58 crc kubenswrapper[5002]: I1014 08:17:58.323661 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" podStartSLOduration=1.9073491420000002 podStartE2EDuration="2.323641998s" podCreationTimestamp="2025-10-14 08:17:56 +0000 UTC" firstStartedPulling="2025-10-14 08:17:57.354957084 +0000 UTC m=+1610.336196536" lastFinishedPulling="2025-10-14 08:17:57.77124994 +0000 UTC m=+1610.752489392" observedRunningTime="2025-10-14 08:17:58.321469459 +0000 UTC m=+1611.302708931" watchObservedRunningTime="2025-10-14 08:17:58.323641998 +0000 UTC m=+1611.304881450" Oct 14 08:18:02 crc kubenswrapper[5002]: I1014 08:18:02.721388 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:18:02 crc kubenswrapper[5002]: E1014 08:18:02.722143 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:18:14 crc kubenswrapper[5002]: I1014 08:18:14.072229 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-7wzf7"] Oct 14 08:18:14 crc kubenswrapper[5002]: I1014 08:18:14.085761 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-7wzf7"] Oct 14 08:18:15 crc kubenswrapper[5002]: I1014 08:18:15.733563 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bbe80a8-92fa-4f55-9937-414e1755a05e" path="/var/lib/kubelet/pods/8bbe80a8-92fa-4f55-9937-414e1755a05e/volumes" Oct 14 08:18:16 crc kubenswrapper[5002]: I1014 08:18:16.721196 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:18:16 crc kubenswrapper[5002]: E1014 08:18:16.721696 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:18:19 crc kubenswrapper[5002]: I1014 08:18:19.033614 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-fbcgt"] Oct 14 08:18:19 crc kubenswrapper[5002]: I1014 08:18:19.041251 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-fbcgt"] Oct 14 08:18:19 crc kubenswrapper[5002]: I1014 08:18:19.735013 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69" path="/var/lib/kubelet/pods/d5131d0b-1f4b-4882-8dc9-ffaaa3a18d69/volumes" Oct 14 08:18:22 crc kubenswrapper[5002]: I1014 08:18:22.760088 5002 scope.go:117] "RemoveContainer" containerID="f9a847a10bd556ae01dc5e8adcc76f5b8bc558d27465de25cb82f9861a0c149e" Oct 14 08:18:22 crc kubenswrapper[5002]: I1014 08:18:22.807555 5002 scope.go:117] "RemoveContainer" containerID="ce85c3c47b4c2af18034f1cf83a1567bb7d189cc076da9d15d5f3ba00bd817b3" Oct 14 08:18:22 crc kubenswrapper[5002]: I1014 08:18:22.878286 5002 scope.go:117] "RemoveContainer" containerID="8f78daae4f3b6f9db13c31b5682eee751191639b981f868ce72155de15da63ac" Oct 14 08:18:22 crc kubenswrapper[5002]: I1014 08:18:22.927197 5002 scope.go:117] "RemoveContainer" containerID="77a7cca8e565256e664d13193a99ac15aafc2b59ed7c2f29a887c5018acc71cd" Oct 14 08:18:22 crc kubenswrapper[5002]: I1014 08:18:22.974477 5002 scope.go:117] "RemoveContainer" containerID="879e0281f9f2c43346dd40061861ddcbe7d0490d0d31eae37489dd2fd5132675" Oct 14 08:18:23 crc kubenswrapper[5002]: I1014 08:18:23.004621 5002 scope.go:117] "RemoveContainer" containerID="2c2aae10edcba2d9bb02e51e4ff61a66699822f5f8ac4830c3ae5669ddb1e5e6" Oct 14 08:18:23 crc kubenswrapper[5002]: I1014 08:18:23.051648 5002 scope.go:117] "RemoveContainer" containerID="efe005d5dad3f56bf17f8f278af3befbf433698afb2da4607320882132b0cbaf" Oct 14 08:18:23 crc kubenswrapper[5002]: I1014 08:18:23.073462 5002 scope.go:117] "RemoveContainer" containerID="7c66370491ebe1acecc02c82a0d9aa51c3410f1027234cc6163a324a8544eca9" Oct 14 08:18:31 crc kubenswrapper[5002]: I1014 08:18:31.721099 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:18:31 crc kubenswrapper[5002]: E1014 08:18:31.722281 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:18:41 crc kubenswrapper[5002]: I1014 08:18:41.055822 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-v746c"] Oct 14 08:18:41 crc kubenswrapper[5002]: I1014 08:18:41.069360 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-v746c"] Oct 14 08:18:41 crc kubenswrapper[5002]: I1014 08:18:41.743396 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0307318a-577d-4b25-9972-274b46d62f36" path="/var/lib/kubelet/pods/0307318a-577d-4b25-9972-274b46d62f36/volumes" Oct 14 08:18:43 crc kubenswrapper[5002]: I1014 08:18:43.721101 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:18:43 crc kubenswrapper[5002]: E1014 08:18:43.721927 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:18:44 crc kubenswrapper[5002]: I1014 08:18:44.049507 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jv98l"] Oct 14 08:18:44 crc kubenswrapper[5002]: I1014 08:18:44.064428 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jv98l"] Oct 14 08:18:45 crc kubenswrapper[5002]: I1014 08:18:45.739150 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce7a28b-9ac5-4a32-a9c9-859df10c79c0" path="/var/lib/kubelet/pods/fce7a28b-9ac5-4a32-a9c9-859df10c79c0/volumes" Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.044075 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-575kw"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.057099 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-c2mx5"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.066342 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-c2mx5"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.075728 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-575kw"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.097706 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4ks2z"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.105964 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4ks2z"] Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.741010 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02a2af3f-d7e5-4e1e-89ba-b567ab162985" path="/var/lib/kubelet/pods/02a2af3f-d7e5-4e1e-89ba-b567ab162985/volumes" Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.742114 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45065e0e-2e90-427d-a49c-423e2ba340d7" path="/var/lib/kubelet/pods/45065e0e-2e90-427d-a49c-423e2ba340d7/volumes" Oct 14 08:18:51 crc kubenswrapper[5002]: I1014 08:18:51.743186 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8" path="/var/lib/kubelet/pods/fb8ffdae-6fd4-461a-8f7b-1f10e06d5dd8/volumes" Oct 14 08:18:56 crc kubenswrapper[5002]: I1014 08:18:56.720267 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:18:56 crc kubenswrapper[5002]: E1014 08:18:56.721395 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:18:57 crc kubenswrapper[5002]: I1014 08:18:57.977061 5002 generic.go:334] "Generic (PLEG): container finished" podID="2fd571d7-b282-49c2-984d-b37b510d0d48" containerID="898c56916e4d9e7bcdb6dbf835e5b43360e691116a7f2ce8289f0e514e58c622" exitCode=2 Oct 14 08:18:57 crc kubenswrapper[5002]: I1014 08:18:57.977184 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" event={"ID":"2fd571d7-b282-49c2-984d-b37b510d0d48","Type":"ContainerDied","Data":"898c56916e4d9e7bcdb6dbf835e5b43360e691116a7f2ce8289f0e514e58c622"} Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.372110 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.496540 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory\") pod \"2fd571d7-b282-49c2-984d-b37b510d0d48\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.496602 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key\") pod \"2fd571d7-b282-49c2-984d-b37b510d0d48\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.496829 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm8jb\" (UniqueName: \"kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb\") pod \"2fd571d7-b282-49c2-984d-b37b510d0d48\" (UID: \"2fd571d7-b282-49c2-984d-b37b510d0d48\") " Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.503403 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb" (OuterVolumeSpecName: "kube-api-access-tm8jb") pod "2fd571d7-b282-49c2-984d-b37b510d0d48" (UID: "2fd571d7-b282-49c2-984d-b37b510d0d48"). InnerVolumeSpecName "kube-api-access-tm8jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.541530 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory" (OuterVolumeSpecName: "inventory") pod "2fd571d7-b282-49c2-984d-b37b510d0d48" (UID: "2fd571d7-b282-49c2-984d-b37b510d0d48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.544528 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fd571d7-b282-49c2-984d-b37b510d0d48" (UID: "2fd571d7-b282-49c2-984d-b37b510d0d48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.599596 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm8jb\" (UniqueName: \"kubernetes.io/projected/2fd571d7-b282-49c2-984d-b37b510d0d48-kube-api-access-tm8jb\") on node \"crc\" DevicePath \"\"" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.599639 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:18:59 crc kubenswrapper[5002]: I1014 08:18:59.599651 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fd571d7-b282-49c2-984d-b37b510d0d48-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:00 crc kubenswrapper[5002]: I1014 08:19:00.005455 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" event={"ID":"2fd571d7-b282-49c2-984d-b37b510d0d48","Type":"ContainerDied","Data":"facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768"} Oct 14 08:19:00 crc kubenswrapper[5002]: I1014 08:19:00.005909 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="facf0b46beccc50dbd24d6e41d2e7c5259944e22a995b0223ddfea3e7930d768" Oct 14 08:19:00 crc kubenswrapper[5002]: I1014 08:19:00.005600 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb" Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.041304 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-886c-account-create-sd5qd"] Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.057224 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9546-account-create-r4d9k"] Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.068535 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-886c-account-create-sd5qd"] Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.078986 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9546-account-create-r4d9k"] Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.740048 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ca6482-5693-4c7d-b49a-81b265362d6b" path="/var/lib/kubelet/pods/d3ca6482-5693-4c7d-b49a-81b265362d6b/volumes" Oct 14 08:19:01 crc kubenswrapper[5002]: I1014 08:19:01.741801 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1037688-f513-41da-9b9b-ae374f5f7732" path="/var/lib/kubelet/pods/e1037688-f513-41da-9b9b-ae374f5f7732/volumes" Oct 14 08:19:02 crc kubenswrapper[5002]: I1014 08:19:02.038927 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2a87-account-create-tmzd4"] Oct 14 08:19:02 crc kubenswrapper[5002]: I1014 08:19:02.048258 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2a87-account-create-tmzd4"] Oct 14 08:19:03 crc kubenswrapper[5002]: I1014 08:19:03.733366 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f066883e-6310-43d3-8805-493589140410" path="/var/lib/kubelet/pods/f066883e-6310-43d3-8805-493589140410/volumes" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.041389 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r"] Oct 14 08:19:07 crc kubenswrapper[5002]: E1014 08:19:07.042559 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd571d7-b282-49c2-984d-b37b510d0d48" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.042593 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd571d7-b282-49c2-984d-b37b510d0d48" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.043138 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd571d7-b282-49c2-984d-b37b510d0d48" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.044402 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.049607 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.050638 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.051697 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.055796 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.060523 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r"] Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.065554 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.065632 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.065743 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjzm2\" (UniqueName: \"kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.167193 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.167279 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.167387 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjzm2\" (UniqueName: \"kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.175510 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.175973 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.188332 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjzm2\" (UniqueName: \"kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2w45r\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.384249 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:19:07 crc kubenswrapper[5002]: I1014 08:19:07.781608 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r"] Oct 14 08:19:08 crc kubenswrapper[5002]: I1014 08:19:08.112461 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" event={"ID":"36c98054-3954-4db3-8f6d-48129782f8a6","Type":"ContainerStarted","Data":"b536f1bb6ae489f577920b41f7729d90ef23fda4911f73a332ae29083bf23ac3"} Oct 14 08:19:08 crc kubenswrapper[5002]: I1014 08:19:08.422206 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:19:09 crc kubenswrapper[5002]: I1014 08:19:09.123423 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" event={"ID":"36c98054-3954-4db3-8f6d-48129782f8a6","Type":"ContainerStarted","Data":"47cc032d29002d819310217291f10845abe0bc7f8eb27ea66bbb34a33dac4303"} Oct 14 08:19:09 crc kubenswrapper[5002]: I1014 08:19:09.150759 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" podStartSLOduration=1.517999472 podStartE2EDuration="2.150738312s" podCreationTimestamp="2025-10-14 08:19:07 +0000 UTC" firstStartedPulling="2025-10-14 08:19:07.786679708 +0000 UTC m=+1680.767919160" lastFinishedPulling="2025-10-14 08:19:08.419418538 +0000 UTC m=+1681.400658000" observedRunningTime="2025-10-14 08:19:09.147230174 +0000 UTC m=+1682.128469686" watchObservedRunningTime="2025-10-14 08:19:09.150738312 +0000 UTC m=+1682.131977754" Oct 14 08:19:11 crc kubenswrapper[5002]: I1014 08:19:11.720377 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:19:11 crc kubenswrapper[5002]: E1014 08:19:11.722917 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.275161 5002 scope.go:117] "RemoveContainer" containerID="d15712662a4a64b0e26d02543cec76a83451768b5bccfc80ab2ccf47811a35d2" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.314379 5002 scope.go:117] "RemoveContainer" containerID="c4b73f9c2443ba98ce446154d4c87a5d39eb6f2bfb951754a550de379f4c2a09" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.400052 5002 scope.go:117] "RemoveContainer" containerID="17431fac5f13d50730a283bf8140e1a7cb93673360b6c9856973c3bfa780fec0" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.438182 5002 scope.go:117] "RemoveContainer" containerID="563a3132c15960a9c0a81ec98db5e8ec19d59aa4551ce458c7f1194127a8629c" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.471035 5002 scope.go:117] "RemoveContainer" containerID="66ac1cb35ce3b20dfde7f1a78effc2cea89c12fdf651b690f8f3283f733d89a2" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.504456 5002 scope.go:117] "RemoveContainer" containerID="05979b86aff0e92996725e7f20034b603345a38e9a4d81b39d9a0d60164377cf" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.564226 5002 scope.go:117] "RemoveContainer" containerID="51d2dd25862e928eb85bb2ca84346923b18e4a123a98b9e16d07a3adbc66de19" Oct 14 08:19:23 crc kubenswrapper[5002]: I1014 08:19:23.583164 5002 scope.go:117] "RemoveContainer" containerID="1610fa5aa92070137909bf008b5d3f51f59c57b77c710f13f193a351ff1444ba" Oct 14 08:19:24 crc kubenswrapper[5002]: I1014 08:19:24.065724 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jrn8p"] Oct 14 08:19:24 crc kubenswrapper[5002]: I1014 08:19:24.080035 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jrn8p"] Oct 14 08:19:25 crc kubenswrapper[5002]: I1014 08:19:25.740559 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad94d62-d1f0-45b9-be90-7bf2b95612a9" path="/var/lib/kubelet/pods/4ad94d62-d1f0-45b9-be90-7bf2b95612a9/volumes" Oct 14 08:19:26 crc kubenswrapper[5002]: I1014 08:19:26.720420 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:19:26 crc kubenswrapper[5002]: E1014 08:19:26.720885 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:19:37 crc kubenswrapper[5002]: I1014 08:19:37.731983 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:19:37 crc kubenswrapper[5002]: E1014 08:19:37.733330 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:19:48 crc kubenswrapper[5002]: I1014 08:19:48.043291 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d5lkh"] Oct 14 08:19:48 crc kubenswrapper[5002]: I1014 08:19:48.051621 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-kxtvt"] Oct 14 08:19:48 crc kubenswrapper[5002]: I1014 08:19:48.059272 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-d5lkh"] Oct 14 08:19:48 crc kubenswrapper[5002]: I1014 08:19:48.070548 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-kxtvt"] Oct 14 08:19:49 crc kubenswrapper[5002]: I1014 08:19:49.721470 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:19:49 crc kubenswrapper[5002]: E1014 08:19:49.722065 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:19:49 crc kubenswrapper[5002]: I1014 08:19:49.740254 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4291140a-1f93-4552-bd10-ccbdb1b17b9f" path="/var/lib/kubelet/pods/4291140a-1f93-4552-bd10-ccbdb1b17b9f/volumes" Oct 14 08:19:49 crc kubenswrapper[5002]: I1014 08:19:49.741796 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="880bc121-fc8a-42e8-bb96-e1c0c615d39e" path="/var/lib/kubelet/pods/880bc121-fc8a-42e8-bb96-e1c0c615d39e/volumes" Oct 14 08:20:02 crc kubenswrapper[5002]: I1014 08:20:02.720814 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:20:02 crc kubenswrapper[5002]: E1014 08:20:02.721471 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:20:02 crc kubenswrapper[5002]: I1014 08:20:02.725233 5002 generic.go:334] "Generic (PLEG): container finished" podID="36c98054-3954-4db3-8f6d-48129782f8a6" containerID="47cc032d29002d819310217291f10845abe0bc7f8eb27ea66bbb34a33dac4303" exitCode=0 Oct 14 08:20:02 crc kubenswrapper[5002]: I1014 08:20:02.725307 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" event={"ID":"36c98054-3954-4db3-8f6d-48129782f8a6","Type":"ContainerDied","Data":"47cc032d29002d819310217291f10845abe0bc7f8eb27ea66bbb34a33dac4303"} Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.174321 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.286102 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjzm2\" (UniqueName: \"kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2\") pod \"36c98054-3954-4db3-8f6d-48129782f8a6\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.288142 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory\") pod \"36c98054-3954-4db3-8f6d-48129782f8a6\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.289359 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key\") pod \"36c98054-3954-4db3-8f6d-48129782f8a6\" (UID: \"36c98054-3954-4db3-8f6d-48129782f8a6\") " Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.300220 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2" (OuterVolumeSpecName: "kube-api-access-kjzm2") pod "36c98054-3954-4db3-8f6d-48129782f8a6" (UID: "36c98054-3954-4db3-8f6d-48129782f8a6"). InnerVolumeSpecName "kube-api-access-kjzm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.324968 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory" (OuterVolumeSpecName: "inventory") pod "36c98054-3954-4db3-8f6d-48129782f8a6" (UID: "36c98054-3954-4db3-8f6d-48129782f8a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.325510 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "36c98054-3954-4db3-8f6d-48129782f8a6" (UID: "36c98054-3954-4db3-8f6d-48129782f8a6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.392532 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.392577 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjzm2\" (UniqueName: \"kubernetes.io/projected/36c98054-3954-4db3-8f6d-48129782f8a6-kube-api-access-kjzm2\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.392597 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36c98054-3954-4db3-8f6d-48129782f8a6-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.748682 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" event={"ID":"36c98054-3954-4db3-8f6d-48129782f8a6","Type":"ContainerDied","Data":"b536f1bb6ae489f577920b41f7729d90ef23fda4911f73a332ae29083bf23ac3"} Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.748740 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b536f1bb6ae489f577920b41f7729d90ef23fda4911f73a332ae29083bf23ac3" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.748828 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.880000 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-czhc8"] Oct 14 08:20:04 crc kubenswrapper[5002]: E1014 08:20:04.880376 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c98054-3954-4db3-8f6d-48129782f8a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.880396 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c98054-3954-4db3-8f6d-48129782f8a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.880599 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c98054-3954-4db3-8f6d-48129782f8a6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.881273 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.887265 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.887269 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.887704 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.887888 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.903005 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.903104 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s45j\" (UniqueName: \"kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.903260 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:04 crc kubenswrapper[5002]: I1014 08:20:04.915474 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-czhc8"] Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.011802 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.011898 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.011940 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s45j\" (UniqueName: \"kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.018790 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.019422 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.041790 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s45j\" (UniqueName: \"kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j\") pod \"ssh-known-hosts-edpm-deployment-czhc8\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.218879 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:05 crc kubenswrapper[5002]: I1014 08:20:05.797009 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-czhc8"] Oct 14 08:20:06 crc kubenswrapper[5002]: I1014 08:20:06.779098 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" event={"ID":"0361ffaa-a5a2-4a5d-b361-a694eb375832","Type":"ContainerStarted","Data":"30e93b2d74e978b9ceb188f29f6921af162b7dee80c52bae5da2fa21cc75df8b"} Oct 14 08:20:06 crc kubenswrapper[5002]: I1014 08:20:06.779458 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" event={"ID":"0361ffaa-a5a2-4a5d-b361-a694eb375832","Type":"ContainerStarted","Data":"b34feb1cb7c3e8b5fcb6b31850a3802ead7552051c459b93648d739c9485aff8"} Oct 14 08:20:06 crc kubenswrapper[5002]: I1014 08:20:06.796777 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" podStartSLOduration=2.290994665 podStartE2EDuration="2.796751343s" podCreationTimestamp="2025-10-14 08:20:04 +0000 UTC" firstStartedPulling="2025-10-14 08:20:05.800934088 +0000 UTC m=+1738.782173550" lastFinishedPulling="2025-10-14 08:20:06.306690736 +0000 UTC m=+1739.287930228" observedRunningTime="2025-10-14 08:20:06.794383116 +0000 UTC m=+1739.775622648" watchObservedRunningTime="2025-10-14 08:20:06.796751343 +0000 UTC m=+1739.777990835" Oct 14 08:20:14 crc kubenswrapper[5002]: I1014 08:20:14.877486 5002 generic.go:334] "Generic (PLEG): container finished" podID="0361ffaa-a5a2-4a5d-b361-a694eb375832" containerID="30e93b2d74e978b9ceb188f29f6921af162b7dee80c52bae5da2fa21cc75df8b" exitCode=0 Oct 14 08:20:14 crc kubenswrapper[5002]: I1014 08:20:14.877621 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" event={"ID":"0361ffaa-a5a2-4a5d-b361-a694eb375832","Type":"ContainerDied","Data":"30e93b2d74e978b9ceb188f29f6921af162b7dee80c52bae5da2fa21cc75df8b"} Oct 14 08:20:15 crc kubenswrapper[5002]: I1014 08:20:15.721658 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:20:15 crc kubenswrapper[5002]: E1014 08:20:15.722801 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.425534 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.461153 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam\") pod \"0361ffaa-a5a2-4a5d-b361-a694eb375832\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.461491 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0\") pod \"0361ffaa-a5a2-4a5d-b361-a694eb375832\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.461700 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s45j\" (UniqueName: \"kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j\") pod \"0361ffaa-a5a2-4a5d-b361-a694eb375832\" (UID: \"0361ffaa-a5a2-4a5d-b361-a694eb375832\") " Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.471406 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j" (OuterVolumeSpecName: "kube-api-access-2s45j") pod "0361ffaa-a5a2-4a5d-b361-a694eb375832" (UID: "0361ffaa-a5a2-4a5d-b361-a694eb375832"). InnerVolumeSpecName "kube-api-access-2s45j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.494021 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0361ffaa-a5a2-4a5d-b361-a694eb375832" (UID: "0361ffaa-a5a2-4a5d-b361-a694eb375832"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.496132 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "0361ffaa-a5a2-4a5d-b361-a694eb375832" (UID: "0361ffaa-a5a2-4a5d-b361-a694eb375832"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.563604 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.563643 5002 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0361ffaa-a5a2-4a5d-b361-a694eb375832-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.563658 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s45j\" (UniqueName: \"kubernetes.io/projected/0361ffaa-a5a2-4a5d-b361-a694eb375832-kube-api-access-2s45j\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.908897 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" event={"ID":"0361ffaa-a5a2-4a5d-b361-a694eb375832","Type":"ContainerDied","Data":"b34feb1cb7c3e8b5fcb6b31850a3802ead7552051c459b93648d739c9485aff8"} Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.908981 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b34feb1cb7c3e8b5fcb6b31850a3802ead7552051c459b93648d739c9485aff8" Oct 14 08:20:16 crc kubenswrapper[5002]: I1014 08:20:16.909014 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-czhc8" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.000702 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5"] Oct 14 08:20:17 crc kubenswrapper[5002]: E1014 08:20:17.001176 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0361ffaa-a5a2-4a5d-b361-a694eb375832" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.001198 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0361ffaa-a5a2-4a5d-b361-a694eb375832" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.001484 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0361ffaa-a5a2-4a5d-b361-a694eb375832" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.002361 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.005187 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.005205 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.005205 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.006789 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.036722 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5"] Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.074503 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.074550 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.074828 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p9df\" (UniqueName: \"kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.175627 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.175683 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.175787 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p9df\" (UniqueName: \"kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.190170 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.209588 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.214882 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p9df\" (UniqueName: \"kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6jsz5\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.364563 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.878229 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5"] Oct 14 08:20:17 crc kubenswrapper[5002]: I1014 08:20:17.916355 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" event={"ID":"46edba04-31ac-4e01-aa14-9b20a4d7ce07","Type":"ContainerStarted","Data":"22c7d0974942e3da9a5df0759bb761a4d4858796fb45784b0e7f1c6f5a0710a8"} Oct 14 08:20:18 crc kubenswrapper[5002]: I1014 08:20:18.925176 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" event={"ID":"46edba04-31ac-4e01-aa14-9b20a4d7ce07","Type":"ContainerStarted","Data":"e6f52024c00739cc014aa9e09b3a5b3f7b63d6f9e2ea49461169afa6753dabb4"} Oct 14 08:20:18 crc kubenswrapper[5002]: I1014 08:20:18.942703 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" podStartSLOduration=2.4052899180000002 podStartE2EDuration="2.942685435s" podCreationTimestamp="2025-10-14 08:20:16 +0000 UTC" firstStartedPulling="2025-10-14 08:20:17.884124366 +0000 UTC m=+1750.865363828" lastFinishedPulling="2025-10-14 08:20:18.421519843 +0000 UTC m=+1751.402759345" observedRunningTime="2025-10-14 08:20:18.941274215 +0000 UTC m=+1751.922513687" watchObservedRunningTime="2025-10-14 08:20:18.942685435 +0000 UTC m=+1751.923924907" Oct 14 08:20:23 crc kubenswrapper[5002]: I1014 08:20:23.725824 5002 scope.go:117] "RemoveContainer" containerID="7bc84491255c430dc9c9c64476decd408acc72d7379ce1a3ed0fba09b11daf08" Oct 14 08:20:23 crc kubenswrapper[5002]: I1014 08:20:23.800342 5002 scope.go:117] "RemoveContainer" containerID="c0960d50784c055f2f1b3fd2f5be65005fdca52cd348b78dd0c07e41fff65106" Oct 14 08:20:23 crc kubenswrapper[5002]: I1014 08:20:23.861925 5002 scope.go:117] "RemoveContainer" containerID="ac189386219a24197085d751764c9963973dc00fbc6a7be042cda9abc8a28dce" Oct 14 08:20:28 crc kubenswrapper[5002]: I1014 08:20:28.018480 5002 generic.go:334] "Generic (PLEG): container finished" podID="46edba04-31ac-4e01-aa14-9b20a4d7ce07" containerID="e6f52024c00739cc014aa9e09b3a5b3f7b63d6f9e2ea49461169afa6753dabb4" exitCode=0 Oct 14 08:20:28 crc kubenswrapper[5002]: I1014 08:20:28.018616 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" event={"ID":"46edba04-31ac-4e01-aa14-9b20a4d7ce07","Type":"ContainerDied","Data":"e6f52024c00739cc014aa9e09b3a5b3f7b63d6f9e2ea49461169afa6753dabb4"} Oct 14 08:20:28 crc kubenswrapper[5002]: I1014 08:20:28.720009 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:20:28 crc kubenswrapper[5002]: E1014 08:20:28.720435 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.518727 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.533697 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key\") pod \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.533794 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p9df\" (UniqueName: \"kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df\") pod \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.533861 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory\") pod \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\" (UID: \"46edba04-31ac-4e01-aa14-9b20a4d7ce07\") " Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.541179 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df" (OuterVolumeSpecName: "kube-api-access-5p9df") pod "46edba04-31ac-4e01-aa14-9b20a4d7ce07" (UID: "46edba04-31ac-4e01-aa14-9b20a4d7ce07"). InnerVolumeSpecName "kube-api-access-5p9df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.567008 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory" (OuterVolumeSpecName: "inventory") pod "46edba04-31ac-4e01-aa14-9b20a4d7ce07" (UID: "46edba04-31ac-4e01-aa14-9b20a4d7ce07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.574984 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46edba04-31ac-4e01-aa14-9b20a4d7ce07" (UID: "46edba04-31ac-4e01-aa14-9b20a4d7ce07"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.637352 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.637411 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p9df\" (UniqueName: \"kubernetes.io/projected/46edba04-31ac-4e01-aa14-9b20a4d7ce07-kube-api-access-5p9df\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:29 crc kubenswrapper[5002]: I1014 08:20:29.637436 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46edba04-31ac-4e01-aa14-9b20a4d7ce07-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.042310 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" event={"ID":"46edba04-31ac-4e01-aa14-9b20a4d7ce07","Type":"ContainerDied","Data":"22c7d0974942e3da9a5df0759bb761a4d4858796fb45784b0e7f1c6f5a0710a8"} Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.042373 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c7d0974942e3da9a5df0759bb761a4d4858796fb45784b0e7f1c6f5a0710a8" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.042401 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.166428 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx"] Oct 14 08:20:30 crc kubenswrapper[5002]: E1014 08:20:30.166974 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46edba04-31ac-4e01-aa14-9b20a4d7ce07" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.166998 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="46edba04-31ac-4e01-aa14-9b20a4d7ce07" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.167220 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="46edba04-31ac-4e01-aa14-9b20a4d7ce07" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.168044 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.174117 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.174374 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.174575 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.177180 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx"] Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.178814 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.252972 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.253561 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.253648 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbtx\" (UniqueName: \"kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.355088 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.355390 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.355475 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbtx\" (UniqueName: \"kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.362756 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.362800 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.382555 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbtx\" (UniqueName: \"kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-klctx\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.505375 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:30 crc kubenswrapper[5002]: I1014 08:20:30.846304 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx"] Oct 14 08:20:31 crc kubenswrapper[5002]: I1014 08:20:31.062012 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" event={"ID":"35271373-a9d9-46f1-92d0-f8301302339f","Type":"ContainerStarted","Data":"8ee412d3a30686ce40421291ab6c4a65732a1e7a67226e002ce29354000b74d3"} Oct 14 08:20:32 crc kubenswrapper[5002]: I1014 08:20:32.063678 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ccnf8"] Oct 14 08:20:32 crc kubenswrapper[5002]: I1014 08:20:32.073140 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" event={"ID":"35271373-a9d9-46f1-92d0-f8301302339f","Type":"ContainerStarted","Data":"af939140c00909621ed6424a2cfe93d1cd3ef6a301bb0b783054e15ac3d2adca"} Oct 14 08:20:32 crc kubenswrapper[5002]: I1014 08:20:32.079873 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ccnf8"] Oct 14 08:20:32 crc kubenswrapper[5002]: I1014 08:20:32.104696 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" podStartSLOduration=1.594752556 podStartE2EDuration="2.10466749s" podCreationTimestamp="2025-10-14 08:20:30 +0000 UTC" firstStartedPulling="2025-10-14 08:20:30.858517184 +0000 UTC m=+1763.839756676" lastFinishedPulling="2025-10-14 08:20:31.368432108 +0000 UTC m=+1764.349671610" observedRunningTime="2025-10-14 08:20:32.098450986 +0000 UTC m=+1765.079690458" watchObservedRunningTime="2025-10-14 08:20:32.10466749 +0000 UTC m=+1765.085906982" Oct 14 08:20:33 crc kubenswrapper[5002]: I1014 08:20:33.743899 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7749993e-625e-4787-9d58-ebdc54edf429" path="/var/lib/kubelet/pods/7749993e-625e-4787-9d58-ebdc54edf429/volumes" Oct 14 08:20:42 crc kubenswrapper[5002]: I1014 08:20:42.204551 5002 generic.go:334] "Generic (PLEG): container finished" podID="35271373-a9d9-46f1-92d0-f8301302339f" containerID="af939140c00909621ed6424a2cfe93d1cd3ef6a301bb0b783054e15ac3d2adca" exitCode=0 Oct 14 08:20:42 crc kubenswrapper[5002]: I1014 08:20:42.204622 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" event={"ID":"35271373-a9d9-46f1-92d0-f8301302339f","Type":"ContainerDied","Data":"af939140c00909621ed6424a2cfe93d1cd3ef6a301bb0b783054e15ac3d2adca"} Oct 14 08:20:42 crc kubenswrapper[5002]: I1014 08:20:42.721730 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.218623 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409"} Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.632271 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.652444 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key\") pod \"35271373-a9d9-46f1-92d0-f8301302339f\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.652569 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory\") pod \"35271373-a9d9-46f1-92d0-f8301302339f\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.652726 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzbtx\" (UniqueName: \"kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx\") pod \"35271373-a9d9-46f1-92d0-f8301302339f\" (UID: \"35271373-a9d9-46f1-92d0-f8301302339f\") " Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.698254 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx" (OuterVolumeSpecName: "kube-api-access-vzbtx") pod "35271373-a9d9-46f1-92d0-f8301302339f" (UID: "35271373-a9d9-46f1-92d0-f8301302339f"). InnerVolumeSpecName "kube-api-access-vzbtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.702289 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory" (OuterVolumeSpecName: "inventory") pod "35271373-a9d9-46f1-92d0-f8301302339f" (UID: "35271373-a9d9-46f1-92d0-f8301302339f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.708026 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35271373-a9d9-46f1-92d0-f8301302339f" (UID: "35271373-a9d9-46f1-92d0-f8301302339f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.754787 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.754999 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzbtx\" (UniqueName: \"kubernetes.io/projected/35271373-a9d9-46f1-92d0-f8301302339f-kube-api-access-vzbtx\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:43 crc kubenswrapper[5002]: I1014 08:20:43.755061 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35271373-a9d9-46f1-92d0-f8301302339f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:20:44 crc kubenswrapper[5002]: I1014 08:20:44.230969 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" event={"ID":"35271373-a9d9-46f1-92d0-f8301302339f","Type":"ContainerDied","Data":"8ee412d3a30686ce40421291ab6c4a65732a1e7a67226e002ce29354000b74d3"} Oct 14 08:20:44 crc kubenswrapper[5002]: I1014 08:20:44.231349 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ee412d3a30686ce40421291ab6c4a65732a1e7a67226e002ce29354000b74d3" Oct 14 08:20:44 crc kubenswrapper[5002]: I1014 08:20:44.231036 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx" Oct 14 08:21:23 crc kubenswrapper[5002]: I1014 08:21:23.970603 5002 scope.go:117] "RemoveContainer" containerID="8784cbc047e8110b041c0f8dc89f0ae2571b6943832959a68cc90e0224393cde" Oct 14 08:23:09 crc kubenswrapper[5002]: I1014 08:23:09.218606 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:23:09 crc kubenswrapper[5002]: I1014 08:23:09.219303 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:23:39 crc kubenswrapper[5002]: I1014 08:23:39.218237 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:23:39 crc kubenswrapper[5002]: I1014 08:23:39.218791 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.218090 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.219140 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.219227 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.220561 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.220693 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409" gracePeriod=600 Oct 14 08:24:09 crc kubenswrapper[5002]: E1014 08:24:09.511949 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f3de631_7681_4103_8510_2b5545bbe8cd.slice/crio-6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f3de631_7681_4103_8510_2b5545bbe8cd.slice/crio-conmon-6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.570512 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409" exitCode=0 Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.570626 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409"} Oct 14 08:24:09 crc kubenswrapper[5002]: I1014 08:24:09.570709 5002 scope.go:117] "RemoveContainer" containerID="a574166b8b5338bb0d1af38e09ac8a80d7097c8688db4e9bf3b3f6aad31a9f63" Oct 14 08:24:10 crc kubenswrapper[5002]: I1014 08:24:10.597159 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08"} Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.711820 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.731163 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.731211 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.740626 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.740655 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.747865 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-zv9s6"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.755203 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qkltt"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.761114 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.766312 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-czhc8"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.772979 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wg8sq"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.778543 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-lxmw9"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.787553 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-kg869"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.792657 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.797752 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.803243 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.808696 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.813823 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-m2l2p"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.819186 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-klctx"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.825366 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr7vb"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.830749 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-czhc8"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.836463 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6jsz5"] Oct 14 08:25:27 crc kubenswrapper[5002]: I1014 08:25:27.841427 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2w45r"] Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.691087 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:29 crc kubenswrapper[5002]: E1014 08:25:29.694616 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35271373-a9d9-46f1-92d0-f8301302339f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.694657 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="35271373-a9d9-46f1-92d0-f8301302339f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.695647 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="35271373-a9d9-46f1-92d0-f8301302339f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.699732 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.717109 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.752326 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0361ffaa-a5a2-4a5d-b361-a694eb375832" path="/var/lib/kubelet/pods/0361ffaa-a5a2-4a5d-b361-a694eb375832/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.753623 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fd571d7-b282-49c2-984d-b37b510d0d48" path="/var/lib/kubelet/pods/2fd571d7-b282-49c2-984d-b37b510d0d48/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.756029 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff80a75-e146-479b-b4b3-74adcc4491fc" path="/var/lib/kubelet/pods/2ff80a75-e146-479b-b4b3-74adcc4491fc/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.758091 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35271373-a9d9-46f1-92d0-f8301302339f" path="/var/lib/kubelet/pods/35271373-a9d9-46f1-92d0-f8301302339f/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.761909 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c98054-3954-4db3-8f6d-48129782f8a6" path="/var/lib/kubelet/pods/36c98054-3954-4db3-8f6d-48129782f8a6/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.764301 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c014d84-bdb3-4e4a-b47f-00e315ed4023" path="/var/lib/kubelet/pods/3c014d84-bdb3-4e4a-b47f-00e315ed4023/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.765684 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46edba04-31ac-4e01-aa14-9b20a4d7ce07" path="/var/lib/kubelet/pods/46edba04-31ac-4e01-aa14-9b20a4d7ce07/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.768007 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="929d05f8-81d3-485e-a16b-50c701235a7a" path="/var/lib/kubelet/pods/929d05f8-81d3-485e-a16b-50c701235a7a/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.768669 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88843ae-ab49-49f1-bc52-f82a2ad016ec" path="/var/lib/kubelet/pods/a88843ae-ab49-49f1-bc52-f82a2ad016ec/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.769384 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be99665c-ce97-4ce5-8e47-2855db76123f" path="/var/lib/kubelet/pods/be99665c-ce97-4ce5-8e47-2855db76123f/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.770746 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f087b0-c7ff-46d2-8134-94f0879ff23c" path="/var/lib/kubelet/pods/e5f087b0-c7ff-46d2-8134-94f0879ff23c/volumes" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.887585 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.888127 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.888509 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.989917 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.990028 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.990153 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.990381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:29 crc kubenswrapper[5002]: I1014 08:25:29.990580 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:30 crc kubenswrapper[5002]: I1014 08:25:30.012683 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz\") pod \"redhat-operators-f4kkm\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:30 crc kubenswrapper[5002]: I1014 08:25:30.048927 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:30 crc kubenswrapper[5002]: I1014 08:25:30.510987 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:31 crc kubenswrapper[5002]: I1014 08:25:31.453523 5002 generic.go:334] "Generic (PLEG): container finished" podID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerID="314ed795e3699c9a221df79b5824afad17bdf645853094d8bf17a9bbff391a19" exitCode=0 Oct 14 08:25:31 crc kubenswrapper[5002]: I1014 08:25:31.453629 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerDied","Data":"314ed795e3699c9a221df79b5824afad17bdf645853094d8bf17a9bbff391a19"} Oct 14 08:25:31 crc kubenswrapper[5002]: I1014 08:25:31.453940 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerStarted","Data":"641c0b7c135f9aa1fe0bc12f04f70743640653aa68110f7619e10e494aad43dd"} Oct 14 08:25:31 crc kubenswrapper[5002]: I1014 08:25:31.456531 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:25:32 crc kubenswrapper[5002]: I1014 08:25:32.481072 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerStarted","Data":"582f08b44196542db23eb70fd6598cad9c0cae58441666688fc09ac98b913ac5"} Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.499994 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2"] Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.502024 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.503921 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.504367 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.504450 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.504503 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.505493 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.513029 5002 generic.go:334] "Generic (PLEG): container finished" podID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerID="582f08b44196542db23eb70fd6598cad9c0cae58441666688fc09ac98b913ac5" exitCode=0 Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.513082 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerDied","Data":"582f08b44196542db23eb70fd6598cad9c0cae58441666688fc09ac98b913ac5"} Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.518455 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2"] Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.665570 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.665650 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.666125 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.666177 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.666401 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkfdx\" (UniqueName: \"kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.768580 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.768679 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkfdx\" (UniqueName: \"kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.768908 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.768992 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.769206 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.775483 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.776563 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.783583 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.785387 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.798442 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkfdx\" (UniqueName: \"kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:33 crc kubenswrapper[5002]: I1014 08:25:33.874640 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:34 crc kubenswrapper[5002]: I1014 08:25:34.424798 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2"] Oct 14 08:25:34 crc kubenswrapper[5002]: I1014 08:25:34.523651 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerStarted","Data":"e00ce0e2d9774eaaa30d8f62e45fcb041832e172c0b8aa9e647cea20256916a8"} Oct 14 08:25:34 crc kubenswrapper[5002]: I1014 08:25:34.525076 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" event={"ID":"fc36bcdf-48eb-4ef8-90c9-eb7842529652","Type":"ContainerStarted","Data":"c38c5a44d0e5815f36a63ab2940ef2926d5c7b0f3582ac988441c7784baac8c4"} Oct 14 08:25:34 crc kubenswrapper[5002]: I1014 08:25:34.549144 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f4kkm" podStartSLOduration=2.884076598 podStartE2EDuration="5.549124329s" podCreationTimestamp="2025-10-14 08:25:29 +0000 UTC" firstStartedPulling="2025-10-14 08:25:31.456149431 +0000 UTC m=+2064.437388913" lastFinishedPulling="2025-10-14 08:25:34.121197172 +0000 UTC m=+2067.102436644" observedRunningTime="2025-10-14 08:25:34.545418941 +0000 UTC m=+2067.526658453" watchObservedRunningTime="2025-10-14 08:25:34.549124329 +0000 UTC m=+2067.530363781" Oct 14 08:25:35 crc kubenswrapper[5002]: I1014 08:25:35.536862 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" event={"ID":"fc36bcdf-48eb-4ef8-90c9-eb7842529652","Type":"ContainerStarted","Data":"4e21a2ef03021a74b5ff28e2e462c35ba3ad9b9ad737d3bb2c38a8c2763d04be"} Oct 14 08:25:35 crc kubenswrapper[5002]: I1014 08:25:35.563978 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" podStartSLOduration=1.833068796 podStartE2EDuration="2.563961082s" podCreationTimestamp="2025-10-14 08:25:33 +0000 UTC" firstStartedPulling="2025-10-14 08:25:34.424761167 +0000 UTC m=+2067.406000619" lastFinishedPulling="2025-10-14 08:25:35.155653453 +0000 UTC m=+2068.136892905" observedRunningTime="2025-10-14 08:25:35.555330805 +0000 UTC m=+2068.536570317" watchObservedRunningTime="2025-10-14 08:25:35.563961082 +0000 UTC m=+2068.545200534" Oct 14 08:25:40 crc kubenswrapper[5002]: I1014 08:25:40.049656 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:40 crc kubenswrapper[5002]: I1014 08:25:40.050085 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:40 crc kubenswrapper[5002]: I1014 08:25:40.130497 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:40 crc kubenswrapper[5002]: I1014 08:25:40.633614 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:40 crc kubenswrapper[5002]: I1014 08:25:40.700144 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:42 crc kubenswrapper[5002]: I1014 08:25:42.602272 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f4kkm" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="registry-server" containerID="cri-o://e00ce0e2d9774eaaa30d8f62e45fcb041832e172c0b8aa9e647cea20256916a8" gracePeriod=2 Oct 14 08:25:44 crc kubenswrapper[5002]: I1014 08:25:44.627532 5002 generic.go:334] "Generic (PLEG): container finished" podID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerID="e00ce0e2d9774eaaa30d8f62e45fcb041832e172c0b8aa9e647cea20256916a8" exitCode=0 Oct 14 08:25:44 crc kubenswrapper[5002]: I1014 08:25:44.628117 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerDied","Data":"e00ce0e2d9774eaaa30d8f62e45fcb041832e172c0b8aa9e647cea20256916a8"} Oct 14 08:25:44 crc kubenswrapper[5002]: I1014 08:25:44.947686 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.110025 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz\") pod \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.110167 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content\") pod \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.110233 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities\") pod \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\" (UID: \"8c8e3897-0d6f-4aa6-a127-d2239217ef23\") " Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.111228 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities" (OuterVolumeSpecName: "utilities") pod "8c8e3897-0d6f-4aa6-a127-d2239217ef23" (UID: "8c8e3897-0d6f-4aa6-a127-d2239217ef23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.122622 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz" (OuterVolumeSpecName: "kube-api-access-fvnsz") pod "8c8e3897-0d6f-4aa6-a127-d2239217ef23" (UID: "8c8e3897-0d6f-4aa6-a127-d2239217ef23"). InnerVolumeSpecName "kube-api-access-fvnsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.191625 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c8e3897-0d6f-4aa6-a127-d2239217ef23" (UID: "8c8e3897-0d6f-4aa6-a127-d2239217ef23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.212810 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.212868 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/8c8e3897-0d6f-4aa6-a127-d2239217ef23-kube-api-access-fvnsz\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.212882 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8e3897-0d6f-4aa6-a127-d2239217ef23-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.645586 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f4kkm" event={"ID":"8c8e3897-0d6f-4aa6-a127-d2239217ef23","Type":"ContainerDied","Data":"641c0b7c135f9aa1fe0bc12f04f70743640653aa68110f7619e10e494aad43dd"} Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.645663 5002 scope.go:117] "RemoveContainer" containerID="e00ce0e2d9774eaaa30d8f62e45fcb041832e172c0b8aa9e647cea20256916a8" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.645710 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f4kkm" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.680428 5002 scope.go:117] "RemoveContainer" containerID="582f08b44196542db23eb70fd6598cad9c0cae58441666688fc09ac98b913ac5" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.708220 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.721211 5002 scope.go:117] "RemoveContainer" containerID="314ed795e3699c9a221df79b5824afad17bdf645853094d8bf17a9bbff391a19" Oct 14 08:25:45 crc kubenswrapper[5002]: I1014 08:25:45.734731 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f4kkm"] Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.193434 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:25:47 crc kubenswrapper[5002]: E1014 08:25:47.194266 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="extract-utilities" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.194285 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="extract-utilities" Oct 14 08:25:47 crc kubenswrapper[5002]: E1014 08:25:47.194301 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="extract-content" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.194310 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="extract-content" Oct 14 08:25:47 crc kubenswrapper[5002]: E1014 08:25:47.194317 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="registry-server" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.194326 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="registry-server" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.194548 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" containerName="registry-server" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.196082 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.213589 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.362396 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.362622 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.362824 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g24vf\" (UniqueName: \"kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.464790 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.464882 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g24vf\" (UniqueName: \"kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.464958 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.465505 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.465533 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.489979 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g24vf\" (UniqueName: \"kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf\") pod \"certified-operators-p8vfv\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.569714 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.676261 5002 generic.go:334] "Generic (PLEG): container finished" podID="fc36bcdf-48eb-4ef8-90c9-eb7842529652" containerID="4e21a2ef03021a74b5ff28e2e462c35ba3ad9b9ad737d3bb2c38a8c2763d04be" exitCode=0 Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.676542 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" event={"ID":"fc36bcdf-48eb-4ef8-90c9-eb7842529652","Type":"ContainerDied","Data":"4e21a2ef03021a74b5ff28e2e462c35ba3ad9b9ad737d3bb2c38a8c2763d04be"} Oct 14 08:25:47 crc kubenswrapper[5002]: I1014 08:25:47.749888 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c8e3897-0d6f-4aa6-a127-d2239217ef23" path="/var/lib/kubelet/pods/8c8e3897-0d6f-4aa6-a127-d2239217ef23/volumes" Oct 14 08:25:48 crc kubenswrapper[5002]: I1014 08:25:48.018393 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:25:48 crc kubenswrapper[5002]: I1014 08:25:48.691024 5002 generic.go:334] "Generic (PLEG): container finished" podID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerID="a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1" exitCode=0 Oct 14 08:25:48 crc kubenswrapper[5002]: I1014 08:25:48.691083 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerDied","Data":"a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1"} Oct 14 08:25:48 crc kubenswrapper[5002]: I1014 08:25:48.691518 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerStarted","Data":"1ae0cd20265dbb161921d99024dc1f781ebad3ec3ad4522d5de6f4d074f9230f"} Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.182172 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.302549 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkfdx\" (UniqueName: \"kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx\") pod \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.302593 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key\") pod \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.302697 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle\") pod \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.302726 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory\") pod \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.302829 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph\") pod \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\" (UID: \"fc36bcdf-48eb-4ef8-90c9-eb7842529652\") " Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.308791 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx" (OuterVolumeSpecName: "kube-api-access-jkfdx") pod "fc36bcdf-48eb-4ef8-90c9-eb7842529652" (UID: "fc36bcdf-48eb-4ef8-90c9-eb7842529652"). InnerVolumeSpecName "kube-api-access-jkfdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.309387 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "fc36bcdf-48eb-4ef8-90c9-eb7842529652" (UID: "fc36bcdf-48eb-4ef8-90c9-eb7842529652"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.310760 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph" (OuterVolumeSpecName: "ceph") pod "fc36bcdf-48eb-4ef8-90c9-eb7842529652" (UID: "fc36bcdf-48eb-4ef8-90c9-eb7842529652"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.353863 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory" (OuterVolumeSpecName: "inventory") pod "fc36bcdf-48eb-4ef8-90c9-eb7842529652" (UID: "fc36bcdf-48eb-4ef8-90c9-eb7842529652"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.359989 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc36bcdf-48eb-4ef8-90c9-eb7842529652" (UID: "fc36bcdf-48eb-4ef8-90c9-eb7842529652"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.405246 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkfdx\" (UniqueName: \"kubernetes.io/projected/fc36bcdf-48eb-4ef8-90c9-eb7842529652-kube-api-access-jkfdx\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.405284 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.405297 5002 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.405337 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.405352 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fc36bcdf-48eb-4ef8-90c9-eb7842529652-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.706753 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" event={"ID":"fc36bcdf-48eb-4ef8-90c9-eb7842529652","Type":"ContainerDied","Data":"c38c5a44d0e5815f36a63ab2940ef2926d5c7b0f3582ac988441c7784baac8c4"} Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.707298 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c38c5a44d0e5815f36a63ab2940ef2926d5c7b0f3582ac988441c7784baac8c4" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.706800 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.712081 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerStarted","Data":"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63"} Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.787286 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk"] Oct 14 08:25:49 crc kubenswrapper[5002]: E1014 08:25:49.787901 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc36bcdf-48eb-4ef8-90c9-eb7842529652" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.787991 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc36bcdf-48eb-4ef8-90c9-eb7842529652" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.788304 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc36bcdf-48eb-4ef8-90c9-eb7842529652" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.789084 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.796958 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk"] Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.822333 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.822574 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.822671 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.822741 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.822680 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.926126 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r7pp\" (UniqueName: \"kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.926186 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.926485 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.926627 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:49 crc kubenswrapper[5002]: I1014 08:25:49.926669 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.028683 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r7pp\" (UniqueName: \"kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.028753 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.028927 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.029029 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.029741 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.035350 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.035551 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.036831 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.037314 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.051371 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r7pp\" (UniqueName: \"kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.137746 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.730573 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk"] Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.732294 5002 generic.go:334] "Generic (PLEG): container finished" podID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerID="0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63" exitCode=0 Oct 14 08:25:50 crc kubenswrapper[5002]: I1014 08:25:50.732359 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerDied","Data":"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63"} Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.594617 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.601933 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.619044 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.670338 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg5gs\" (UniqueName: \"kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.670589 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.670818 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.741825 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerStarted","Data":"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b"} Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.743709 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" event={"ID":"938c7955-b34d-435c-8933-2c3dc8043578","Type":"ContainerStarted","Data":"3e5ab507ba3da906882eba085a1a4f4ceb5509323873a7f20d4d68b5a7d3b5dd"} Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.743735 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" event={"ID":"938c7955-b34d-435c-8933-2c3dc8043578","Type":"ContainerStarted","Data":"980f0b92f3129c466793237eca11ee17157659be3c5324043f08df4a580ef0d3"} Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.772283 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.772378 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.772421 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg5gs\" (UniqueName: \"kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.773236 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.773330 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.775529 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p8vfv" podStartSLOduration=2.340359754 podStartE2EDuration="4.775511369s" podCreationTimestamp="2025-10-14 08:25:47 +0000 UTC" firstStartedPulling="2025-10-14 08:25:48.693068548 +0000 UTC m=+2081.674308020" lastFinishedPulling="2025-10-14 08:25:51.128220143 +0000 UTC m=+2084.109459635" observedRunningTime="2025-10-14 08:25:51.769498751 +0000 UTC m=+2084.750738203" watchObservedRunningTime="2025-10-14 08:25:51.775511369 +0000 UTC m=+2084.756750841" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.790934 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" podStartSLOduration=2.330866713 podStartE2EDuration="2.790911654s" podCreationTimestamp="2025-10-14 08:25:49 +0000 UTC" firstStartedPulling="2025-10-14 08:25:50.75250814 +0000 UTC m=+2083.733747662" lastFinishedPulling="2025-10-14 08:25:51.212553111 +0000 UTC m=+2084.193792603" observedRunningTime="2025-10-14 08:25:51.781905027 +0000 UTC m=+2084.763144489" watchObservedRunningTime="2025-10-14 08:25:51.790911654 +0000 UTC m=+2084.772151126" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.804193 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg5gs\" (UniqueName: \"kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs\") pod \"community-operators-q9n9f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:51 crc kubenswrapper[5002]: I1014 08:25:51.926045 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:25:52 crc kubenswrapper[5002]: I1014 08:25:52.510650 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:25:52 crc kubenswrapper[5002]: I1014 08:25:52.753785 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerStarted","Data":"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722"} Oct 14 08:25:52 crc kubenswrapper[5002]: I1014 08:25:52.753853 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerStarted","Data":"1204531e72e4dbcdc9dbac1db058737e988935389cc952a1c758ae6708a35cc2"} Oct 14 08:25:53 crc kubenswrapper[5002]: I1014 08:25:53.764889 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerID="d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722" exitCode=0 Oct 14 08:25:53 crc kubenswrapper[5002]: I1014 08:25:53.764955 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerDied","Data":"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722"} Oct 14 08:25:54 crc kubenswrapper[5002]: I1014 08:25:54.780208 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerStarted","Data":"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41"} Oct 14 08:25:55 crc kubenswrapper[5002]: I1014 08:25:55.796417 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerID="a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41" exitCode=0 Oct 14 08:25:55 crc kubenswrapper[5002]: I1014 08:25:55.796483 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerDied","Data":"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41"} Oct 14 08:25:56 crc kubenswrapper[5002]: I1014 08:25:56.813025 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerStarted","Data":"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff"} Oct 14 08:25:56 crc kubenswrapper[5002]: I1014 08:25:56.846639 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q9n9f" podStartSLOduration=3.297284332 podStartE2EDuration="5.846612729s" podCreationTimestamp="2025-10-14 08:25:51 +0000 UTC" firstStartedPulling="2025-10-14 08:25:53.767199278 +0000 UTC m=+2086.748438770" lastFinishedPulling="2025-10-14 08:25:56.316527675 +0000 UTC m=+2089.297767167" observedRunningTime="2025-10-14 08:25:56.837012346 +0000 UTC m=+2089.818251858" watchObservedRunningTime="2025-10-14 08:25:56.846612729 +0000 UTC m=+2089.827852221" Oct 14 08:25:57 crc kubenswrapper[5002]: I1014 08:25:57.627073 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:57 crc kubenswrapper[5002]: I1014 08:25:57.628302 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:57 crc kubenswrapper[5002]: I1014 08:25:57.676654 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:57 crc kubenswrapper[5002]: I1014 08:25:57.892476 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:25:59 crc kubenswrapper[5002]: I1014 08:25:59.591267 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:26:00 crc kubenswrapper[5002]: I1014 08:26:00.857119 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p8vfv" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="registry-server" containerID="cri-o://4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b" gracePeriod=2 Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.406231 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.536800 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content\") pod \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.536953 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g24vf\" (UniqueName: \"kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf\") pod \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.537051 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities\") pod \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\" (UID: \"1bdefb34-cc12-4fda-a1db-31925a3ea0f5\") " Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.538811 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities" (OuterVolumeSpecName: "utilities") pod "1bdefb34-cc12-4fda-a1db-31925a3ea0f5" (UID: "1bdefb34-cc12-4fda-a1db-31925a3ea0f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.543438 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf" (OuterVolumeSpecName: "kube-api-access-g24vf") pod "1bdefb34-cc12-4fda-a1db-31925a3ea0f5" (UID: "1bdefb34-cc12-4fda-a1db-31925a3ea0f5"). InnerVolumeSpecName "kube-api-access-g24vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.609195 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bdefb34-cc12-4fda-a1db-31925a3ea0f5" (UID: "1bdefb34-cc12-4fda-a1db-31925a3ea0f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.638928 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.639001 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.639023 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g24vf\" (UniqueName: \"kubernetes.io/projected/1bdefb34-cc12-4fda-a1db-31925a3ea0f5-kube-api-access-g24vf\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.880390 5002 generic.go:334] "Generic (PLEG): container finished" podID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerID="4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b" exitCode=0 Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.880441 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerDied","Data":"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b"} Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.880477 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8vfv" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.880497 5002 scope.go:117] "RemoveContainer" containerID="4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.880482 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8vfv" event={"ID":"1bdefb34-cc12-4fda-a1db-31925a3ea0f5","Type":"ContainerDied","Data":"1ae0cd20265dbb161921d99024dc1f781ebad3ec3ad4522d5de6f4d074f9230f"} Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.921264 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.928883 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.928943 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.931019 5002 scope.go:117] "RemoveContainer" containerID="0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63" Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.935173 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p8vfv"] Oct 14 08:26:01 crc kubenswrapper[5002]: I1014 08:26:01.967677 5002 scope.go:117] "RemoveContainer" containerID="a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.009999 5002 scope.go:117] "RemoveContainer" containerID="4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b" Oct 14 08:26:02 crc kubenswrapper[5002]: E1014 08:26:02.010805 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b\": container with ID starting with 4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b not found: ID does not exist" containerID="4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.010844 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b"} err="failed to get container status \"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b\": rpc error: code = NotFound desc = could not find container \"4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b\": container with ID starting with 4c8754b3121232190efeedb32d4fb6f2b562fd3967c47b80958ccfd25d2eb30b not found: ID does not exist" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.010883 5002 scope.go:117] "RemoveContainer" containerID="0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63" Oct 14 08:26:02 crc kubenswrapper[5002]: E1014 08:26:02.011184 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63\": container with ID starting with 0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63 not found: ID does not exist" containerID="0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.011209 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63"} err="failed to get container status \"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63\": rpc error: code = NotFound desc = could not find container \"0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63\": container with ID starting with 0f8d30e0a1f92aed90d892e8ccc808e35f497457acae6154c26d33692c163c63 not found: ID does not exist" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.011226 5002 scope.go:117] "RemoveContainer" containerID="a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1" Oct 14 08:26:02 crc kubenswrapper[5002]: E1014 08:26:02.011552 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1\": container with ID starting with a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1 not found: ID does not exist" containerID="a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.011579 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1"} err="failed to get container status \"a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1\": rpc error: code = NotFound desc = could not find container \"a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1\": container with ID starting with a1f366a9ed004ff3c4dbe6080908e714f02c896ecd9608859db47c2360f4cec1 not found: ID does not exist" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.015498 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:02 crc kubenswrapper[5002]: I1014 08:26:02.943050 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:03 crc kubenswrapper[5002]: I1014 08:26:03.733183 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" path="/var/lib/kubelet/pods/1bdefb34-cc12-4fda-a1db-31925a3ea0f5/volumes" Oct 14 08:26:04 crc kubenswrapper[5002]: I1014 08:26:04.393606 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:26:04 crc kubenswrapper[5002]: I1014 08:26:04.914511 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q9n9f" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="registry-server" containerID="cri-o://27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff" gracePeriod=2 Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.400254 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.519133 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities\") pod \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.519456 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg5gs\" (UniqueName: \"kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs\") pod \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.519612 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content\") pod \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\" (UID: \"cf26a795-9351-4285-8ddf-32bdf7e5f95f\") " Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.519953 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities" (OuterVolumeSpecName: "utilities") pod "cf26a795-9351-4285-8ddf-32bdf7e5f95f" (UID: "cf26a795-9351-4285-8ddf-32bdf7e5f95f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.520571 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.524170 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs" (OuterVolumeSpecName: "kube-api-access-tg5gs") pod "cf26a795-9351-4285-8ddf-32bdf7e5f95f" (UID: "cf26a795-9351-4285-8ddf-32bdf7e5f95f"). InnerVolumeSpecName "kube-api-access-tg5gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.570904 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf26a795-9351-4285-8ddf-32bdf7e5f95f" (UID: "cf26a795-9351-4285-8ddf-32bdf7e5f95f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.622268 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg5gs\" (UniqueName: \"kubernetes.io/projected/cf26a795-9351-4285-8ddf-32bdf7e5f95f-kube-api-access-tg5gs\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.622307 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf26a795-9351-4285-8ddf-32bdf7e5f95f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.927351 5002 generic.go:334] "Generic (PLEG): container finished" podID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerID="27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff" exitCode=0 Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.927390 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerDied","Data":"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff"} Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.927417 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9n9f" event={"ID":"cf26a795-9351-4285-8ddf-32bdf7e5f95f","Type":"ContainerDied","Data":"1204531e72e4dbcdc9dbac1db058737e988935389cc952a1c758ae6708a35cc2"} Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.927432 5002 scope.go:117] "RemoveContainer" containerID="27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.927563 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9n9f" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.958662 5002 scope.go:117] "RemoveContainer" containerID="a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.968595 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.980543 5002 scope.go:117] "RemoveContainer" containerID="d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722" Oct 14 08:26:05 crc kubenswrapper[5002]: I1014 08:26:05.985626 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q9n9f"] Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.035225 5002 scope.go:117] "RemoveContainer" containerID="27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff" Oct 14 08:26:06 crc kubenswrapper[5002]: E1014 08:26:06.035701 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff\": container with ID starting with 27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff not found: ID does not exist" containerID="27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff" Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.035744 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff"} err="failed to get container status \"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff\": rpc error: code = NotFound desc = could not find container \"27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff\": container with ID starting with 27559c6f8044fb70dc991674576415e9cf774477f881d7a9d9181e37671f2dff not found: ID does not exist" Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.035770 5002 scope.go:117] "RemoveContainer" containerID="a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41" Oct 14 08:26:06 crc kubenswrapper[5002]: E1014 08:26:06.036385 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41\": container with ID starting with a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41 not found: ID does not exist" containerID="a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41" Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.036454 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41"} err="failed to get container status \"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41\": rpc error: code = NotFound desc = could not find container \"a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41\": container with ID starting with a4223029736e48ea1d28d49701838aaa5132ef223b8bc19e08679f2bca8bac41 not found: ID does not exist" Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.036499 5002 scope.go:117] "RemoveContainer" containerID="d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722" Oct 14 08:26:06 crc kubenswrapper[5002]: E1014 08:26:06.037028 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722\": container with ID starting with d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722 not found: ID does not exist" containerID="d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722" Oct 14 08:26:06 crc kubenswrapper[5002]: I1014 08:26:06.037079 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722"} err="failed to get container status \"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722\": rpc error: code = NotFound desc = could not find container \"d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722\": container with ID starting with d0163d5d046d986436c87165c2be68e1e8229aa06c73ced0282d033fb97c9722 not found: ID does not exist" Oct 14 08:26:07 crc kubenswrapper[5002]: I1014 08:26:07.735246 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" path="/var/lib/kubelet/pods/cf26a795-9351-4285-8ddf-32bdf7e5f95f/volumes" Oct 14 08:26:09 crc kubenswrapper[5002]: I1014 08:26:09.218690 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:26:09 crc kubenswrapper[5002]: I1014 08:26:09.219078 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.180510 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181496 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181512 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181525 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="extract-content" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181532 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="extract-content" Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181565 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="extract-utilities" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181572 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="extract-utilities" Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181598 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="extract-utilities" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181606 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="extract-utilities" Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181620 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181627 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: E1014 08:26:23.181639 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="extract-content" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181646 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="extract-content" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181875 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdefb34-cc12-4fda-a1db-31925a3ea0f5" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.181903 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf26a795-9351-4285-8ddf-32bdf7e5f95f" containerName="registry-server" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.183480 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.191406 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.365037 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.365542 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.365647 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjhrq\" (UniqueName: \"kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.467613 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.468138 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.468631 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.468658 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjhrq\" (UniqueName: \"kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.468917 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.497225 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjhrq\" (UniqueName: \"kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq\") pod \"redhat-marketplace-n622x\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:23 crc kubenswrapper[5002]: I1014 08:26:23.512044 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.031129 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.134889 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerStarted","Data":"8ee9624e6fbd805095949e93cd1d946dbb17b83b82d22eda4fa181aa9f37215c"} Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.172739 5002 scope.go:117] "RemoveContainer" containerID="7c953250b86cad42d7e9267f8d5627e61dff0d0e304dcde3407698eb4bc111a1" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.213234 5002 scope.go:117] "RemoveContainer" containerID="30e93b2d74e978b9ceb188f29f6921af162b7dee80c52bae5da2fa21cc75df8b" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.254016 5002 scope.go:117] "RemoveContainer" containerID="47cc032d29002d819310217291f10845abe0bc7f8eb27ea66bbb34a33dac4303" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.296889 5002 scope.go:117] "RemoveContainer" containerID="3371c2d98c7c31e767da693319499d2d3528e35a8c95219bf4615f157bdd9c81" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.407554 5002 scope.go:117] "RemoveContainer" containerID="898c56916e4d9e7bcdb6dbf835e5b43360e691116a7f2ce8289f0e514e58c622" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.462634 5002 scope.go:117] "RemoveContainer" containerID="20850f97df7ef558d1ba5adfc9b22d26cf093d495225451dcce1629b566540f5" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.534866 5002 scope.go:117] "RemoveContainer" containerID="ed220500ec6be424d67c9d0da1600496d23405619f8f1757e321531c68117089" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.579352 5002 scope.go:117] "RemoveContainer" containerID="39ccaaafac8f404b9ba7b864a496f9664eda9132ba4eae1ab74f219f99aef7c7" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.609165 5002 scope.go:117] "RemoveContainer" containerID="0af358cc8b22a0d13a188104cee46c4e8c77e58bdc3862f7bacaeee066260fc3" Oct 14 08:26:24 crc kubenswrapper[5002]: I1014 08:26:24.652236 5002 scope.go:117] "RemoveContainer" containerID="e6f52024c00739cc014aa9e09b3a5b3f7b63d6f9e2ea49461169afa6753dabb4" Oct 14 08:26:25 crc kubenswrapper[5002]: I1014 08:26:25.147353 5002 generic.go:334] "Generic (PLEG): container finished" podID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerID="e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4" exitCode=0 Oct 14 08:26:25 crc kubenswrapper[5002]: I1014 08:26:25.147425 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerDied","Data":"e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4"} Oct 14 08:26:27 crc kubenswrapper[5002]: I1014 08:26:27.174934 5002 generic.go:334] "Generic (PLEG): container finished" podID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerID="6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c" exitCode=0 Oct 14 08:26:27 crc kubenswrapper[5002]: I1014 08:26:27.175067 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerDied","Data":"6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c"} Oct 14 08:26:28 crc kubenswrapper[5002]: I1014 08:26:28.187573 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerStarted","Data":"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7"} Oct 14 08:26:28 crc kubenswrapper[5002]: I1014 08:26:28.213753 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n622x" podStartSLOduration=2.518632711 podStartE2EDuration="5.213731943s" podCreationTimestamp="2025-10-14 08:26:23 +0000 UTC" firstStartedPulling="2025-10-14 08:26:25.149570154 +0000 UTC m=+2118.130809606" lastFinishedPulling="2025-10-14 08:26:27.844669386 +0000 UTC m=+2120.825908838" observedRunningTime="2025-10-14 08:26:28.206814842 +0000 UTC m=+2121.188054324" watchObservedRunningTime="2025-10-14 08:26:28.213731943 +0000 UTC m=+2121.194971405" Oct 14 08:26:33 crc kubenswrapper[5002]: I1014 08:26:33.512922 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:33 crc kubenswrapper[5002]: I1014 08:26:33.513686 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:33 crc kubenswrapper[5002]: I1014 08:26:33.594081 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:34 crc kubenswrapper[5002]: I1014 08:26:34.347054 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:34 crc kubenswrapper[5002]: I1014 08:26:34.412243 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.285025 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n622x" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="registry-server" containerID="cri-o://b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7" gracePeriod=2 Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.845710 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.946176 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content\") pod \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.946268 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjhrq\" (UniqueName: \"kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq\") pod \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.946583 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities\") pod \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\" (UID: \"20ea5a76-f152-4f01-8cc2-bb5f2a66df85\") " Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.949347 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities" (OuterVolumeSpecName: "utilities") pod "20ea5a76-f152-4f01-8cc2-bb5f2a66df85" (UID: "20ea5a76-f152-4f01-8cc2-bb5f2a66df85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.953655 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq" (OuterVolumeSpecName: "kube-api-access-zjhrq") pod "20ea5a76-f152-4f01-8cc2-bb5f2a66df85" (UID: "20ea5a76-f152-4f01-8cc2-bb5f2a66df85"). InnerVolumeSpecName "kube-api-access-zjhrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:26:36 crc kubenswrapper[5002]: I1014 08:26:36.959864 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20ea5a76-f152-4f01-8cc2-bb5f2a66df85" (UID: "20ea5a76-f152-4f01-8cc2-bb5f2a66df85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.049310 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.049355 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjhrq\" (UniqueName: \"kubernetes.io/projected/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-kube-api-access-zjhrq\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.049378 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20ea5a76-f152-4f01-8cc2-bb5f2a66df85-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.300477 5002 generic.go:334] "Generic (PLEG): container finished" podID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerID="b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7" exitCode=0 Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.300610 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n622x" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.300598 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerDied","Data":"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7"} Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.301119 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n622x" event={"ID":"20ea5a76-f152-4f01-8cc2-bb5f2a66df85","Type":"ContainerDied","Data":"8ee9624e6fbd805095949e93cd1d946dbb17b83b82d22eda4fa181aa9f37215c"} Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.301151 5002 scope.go:117] "RemoveContainer" containerID="b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.327630 5002 scope.go:117] "RemoveContainer" containerID="6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.355850 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.366466 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n622x"] Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.370802 5002 scope.go:117] "RemoveContainer" containerID="e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.397281 5002 scope.go:117] "RemoveContainer" containerID="b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7" Oct 14 08:26:37 crc kubenswrapper[5002]: E1014 08:26:37.397708 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7\": container with ID starting with b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7 not found: ID does not exist" containerID="b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.397753 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7"} err="failed to get container status \"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7\": rpc error: code = NotFound desc = could not find container \"b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7\": container with ID starting with b87f770f78ae6608c4a184860f53d46850a1aba560c4398c2db108603b19cfa7 not found: ID does not exist" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.397783 5002 scope.go:117] "RemoveContainer" containerID="6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c" Oct 14 08:26:37 crc kubenswrapper[5002]: E1014 08:26:37.398043 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c\": container with ID starting with 6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c not found: ID does not exist" containerID="6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.398072 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c"} err="failed to get container status \"6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c\": rpc error: code = NotFound desc = could not find container \"6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c\": container with ID starting with 6bab816d6c89406c282e9706aadf65d44d8cb03d1292ead212ba5fffb8de533c not found: ID does not exist" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.398088 5002 scope.go:117] "RemoveContainer" containerID="e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4" Oct 14 08:26:37 crc kubenswrapper[5002]: E1014 08:26:37.398289 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4\": container with ID starting with e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4 not found: ID does not exist" containerID="e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.398310 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4"} err="failed to get container status \"e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4\": rpc error: code = NotFound desc = could not find container \"e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4\": container with ID starting with e07dcd0587ff8358715be33fa383f8b48d0019653fb86655d1b7bc5c399f34c4 not found: ID does not exist" Oct 14 08:26:37 crc kubenswrapper[5002]: I1014 08:26:37.732952 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" path="/var/lib/kubelet/pods/20ea5a76-f152-4f01-8cc2-bb5f2a66df85/volumes" Oct 14 08:26:39 crc kubenswrapper[5002]: I1014 08:26:39.218475 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:26:39 crc kubenswrapper[5002]: I1014 08:26:39.219039 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.218253 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.218759 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.218797 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.219485 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.219555 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" gracePeriod=600 Oct 14 08:27:09 crc kubenswrapper[5002]: E1014 08:27:09.349815 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.683954 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" exitCode=0 Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.684021 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08"} Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.684767 5002 scope.go:117] "RemoveContainer" containerID="6223920fb17b6641682444366b701fbdfadfe89af1642daacba4bdb41004f409" Oct 14 08:27:09 crc kubenswrapper[5002]: I1014 08:27:09.686164 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:27:09 crc kubenswrapper[5002]: E1014 08:27:09.686948 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:27:24 crc kubenswrapper[5002]: I1014 08:27:24.721715 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:27:24 crc kubenswrapper[5002]: E1014 08:27:24.722645 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:27:24 crc kubenswrapper[5002]: I1014 08:27:24.938179 5002 scope.go:117] "RemoveContainer" containerID="af939140c00909621ed6424a2cfe93d1cd3ef6a301bb0b783054e15ac3d2adca" Oct 14 08:27:31 crc kubenswrapper[5002]: I1014 08:27:31.925438 5002 generic.go:334] "Generic (PLEG): container finished" podID="938c7955-b34d-435c-8933-2c3dc8043578" containerID="3e5ab507ba3da906882eba085a1a4f4ceb5509323873a7f20d4d68b5a7d3b5dd" exitCode=0 Oct 14 08:27:31 crc kubenswrapper[5002]: I1014 08:27:31.925565 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" event={"ID":"938c7955-b34d-435c-8933-2c3dc8043578","Type":"ContainerDied","Data":"3e5ab507ba3da906882eba085a1a4f4ceb5509323873a7f20d4d68b5a7d3b5dd"} Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.392996 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.490295 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory\") pod \"938c7955-b34d-435c-8933-2c3dc8043578\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.490449 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r7pp\" (UniqueName: \"kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp\") pod \"938c7955-b34d-435c-8933-2c3dc8043578\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.490509 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key\") pod \"938c7955-b34d-435c-8933-2c3dc8043578\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.490532 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle\") pod \"938c7955-b34d-435c-8933-2c3dc8043578\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.490586 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph\") pod \"938c7955-b34d-435c-8933-2c3dc8043578\" (UID: \"938c7955-b34d-435c-8933-2c3dc8043578\") " Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.497142 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "938c7955-b34d-435c-8933-2c3dc8043578" (UID: "938c7955-b34d-435c-8933-2c3dc8043578"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.497646 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp" (OuterVolumeSpecName: "kube-api-access-9r7pp") pod "938c7955-b34d-435c-8933-2c3dc8043578" (UID: "938c7955-b34d-435c-8933-2c3dc8043578"). InnerVolumeSpecName "kube-api-access-9r7pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.497856 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph" (OuterVolumeSpecName: "ceph") pod "938c7955-b34d-435c-8933-2c3dc8043578" (UID: "938c7955-b34d-435c-8933-2c3dc8043578"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.518100 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory" (OuterVolumeSpecName: "inventory") pod "938c7955-b34d-435c-8933-2c3dc8043578" (UID: "938c7955-b34d-435c-8933-2c3dc8043578"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.523002 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "938c7955-b34d-435c-8933-2c3dc8043578" (UID: "938c7955-b34d-435c-8933-2c3dc8043578"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.592934 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.592992 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r7pp\" (UniqueName: \"kubernetes.io/projected/938c7955-b34d-435c-8933-2c3dc8043578-kube-api-access-9r7pp\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.593015 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.593036 5002 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.593055 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/938c7955-b34d-435c-8933-2c3dc8043578-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.948089 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" event={"ID":"938c7955-b34d-435c-8933-2c3dc8043578","Type":"ContainerDied","Data":"980f0b92f3129c466793237eca11ee17157659be3c5324043f08df4a580ef0d3"} Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.948138 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="980f0b92f3129c466793237eca11ee17157659be3c5324043f08df4a580ef0d3" Oct 14 08:27:33 crc kubenswrapper[5002]: I1014 08:27:33.948179 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.061300 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8"] Oct 14 08:27:34 crc kubenswrapper[5002]: E1014 08:27:34.061690 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938c7955-b34d-435c-8933-2c3dc8043578" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.061711 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="938c7955-b34d-435c-8933-2c3dc8043578" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:27:34 crc kubenswrapper[5002]: E1014 08:27:34.061736 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="registry-server" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.061745 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="registry-server" Oct 14 08:27:34 crc kubenswrapper[5002]: E1014 08:27:34.061774 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="extract-utilities" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.061782 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="extract-utilities" Oct 14 08:27:34 crc kubenswrapper[5002]: E1014 08:27:34.061814 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="extract-content" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.061823 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="extract-content" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.062042 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ea5a76-f152-4f01-8cc2-bb5f2a66df85" containerName="registry-server" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.062076 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="938c7955-b34d-435c-8933-2c3dc8043578" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.062763 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.067434 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.068150 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.068374 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.068512 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.068549 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.079240 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8"] Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.205621 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.205686 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.205878 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.205958 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d962p\" (UniqueName: \"kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.307696 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.307792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d962p\" (UniqueName: \"kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.307930 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.307968 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.312700 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.313222 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.313237 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.336499 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d962p\" (UniqueName: \"kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.403806 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:27:34 crc kubenswrapper[5002]: I1014 08:27:34.994774 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8"] Oct 14 08:27:34 crc kubenswrapper[5002]: W1014 08:27:34.998133 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6cb9d07_96cd_42d8_91a1_230e4f6316f0.slice/crio-ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79 WatchSource:0}: Error finding container ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79: Status 404 returned error can't find the container with id ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79 Oct 14 08:27:35 crc kubenswrapper[5002]: I1014 08:27:35.720601 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:27:35 crc kubenswrapper[5002]: E1014 08:27:35.721199 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:27:35 crc kubenswrapper[5002]: I1014 08:27:35.997494 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" event={"ID":"a6cb9d07-96cd-42d8-91a1-230e4f6316f0","Type":"ContainerStarted","Data":"7b268c9133461c7f8fdc52d90d232711d754fdb08bfb021c08290b09c775de84"} Oct 14 08:27:35 crc kubenswrapper[5002]: I1014 08:27:35.998120 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" event={"ID":"a6cb9d07-96cd-42d8-91a1-230e4f6316f0","Type":"ContainerStarted","Data":"ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79"} Oct 14 08:27:36 crc kubenswrapper[5002]: I1014 08:27:36.016879 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" podStartSLOduration=1.416813927 podStartE2EDuration="2.016861312s" podCreationTimestamp="2025-10-14 08:27:34 +0000 UTC" firstStartedPulling="2025-10-14 08:27:35.006451783 +0000 UTC m=+2187.987691235" lastFinishedPulling="2025-10-14 08:27:35.606499148 +0000 UTC m=+2188.587738620" observedRunningTime="2025-10-14 08:27:36.015474945 +0000 UTC m=+2188.996714407" watchObservedRunningTime="2025-10-14 08:27:36.016861312 +0000 UTC m=+2188.998100784" Oct 14 08:27:46 crc kubenswrapper[5002]: I1014 08:27:46.720934 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:27:46 crc kubenswrapper[5002]: E1014 08:27:46.723308 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:27:59 crc kubenswrapper[5002]: I1014 08:27:59.720948 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:27:59 crc kubenswrapper[5002]: E1014 08:27:59.721610 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:28:04 crc kubenswrapper[5002]: I1014 08:28:04.311631 5002 generic.go:334] "Generic (PLEG): container finished" podID="a6cb9d07-96cd-42d8-91a1-230e4f6316f0" containerID="7b268c9133461c7f8fdc52d90d232711d754fdb08bfb021c08290b09c775de84" exitCode=0 Oct 14 08:28:04 crc kubenswrapper[5002]: I1014 08:28:04.311761 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" event={"ID":"a6cb9d07-96cd-42d8-91a1-230e4f6316f0","Type":"ContainerDied","Data":"7b268c9133461c7f8fdc52d90d232711d754fdb08bfb021c08290b09c775de84"} Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.763174 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.867266 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory\") pod \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.867328 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph\") pod \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.867411 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key\") pod \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.867475 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d962p\" (UniqueName: \"kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p\") pod \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\" (UID: \"a6cb9d07-96cd-42d8-91a1-230e4f6316f0\") " Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.887188 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p" (OuterVolumeSpecName: "kube-api-access-d962p") pod "a6cb9d07-96cd-42d8-91a1-230e4f6316f0" (UID: "a6cb9d07-96cd-42d8-91a1-230e4f6316f0"). InnerVolumeSpecName "kube-api-access-d962p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.895493 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a6cb9d07-96cd-42d8-91a1-230e4f6316f0" (UID: "a6cb9d07-96cd-42d8-91a1-230e4f6316f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.902208 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph" (OuterVolumeSpecName: "ceph") pod "a6cb9d07-96cd-42d8-91a1-230e4f6316f0" (UID: "a6cb9d07-96cd-42d8-91a1-230e4f6316f0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.915974 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory" (OuterVolumeSpecName: "inventory") pod "a6cb9d07-96cd-42d8-91a1-230e4f6316f0" (UID: "a6cb9d07-96cd-42d8-91a1-230e4f6316f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.969168 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.969209 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.969217 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:05 crc kubenswrapper[5002]: I1014 08:28:05.969228 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d962p\" (UniqueName: \"kubernetes.io/projected/a6cb9d07-96cd-42d8-91a1-230e4f6316f0-kube-api-access-d962p\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.333396 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" event={"ID":"a6cb9d07-96cd-42d8-91a1-230e4f6316f0","Type":"ContainerDied","Data":"ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79"} Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.333695 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef9274e96af250a7b6e3d356d9a6370926e921b548e692d3595dde71aa0fbf79" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.333507 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.436180 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn"] Oct 14 08:28:06 crc kubenswrapper[5002]: E1014 08:28:06.436658 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cb9d07-96cd-42d8-91a1-230e4f6316f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.436682 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cb9d07-96cd-42d8-91a1-230e4f6316f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.436990 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cb9d07-96cd-42d8-91a1-230e4f6316f0" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.437924 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.441568 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.442922 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.443014 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.444732 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.445179 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.456306 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn"] Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.579861 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrwf\" (UniqueName: \"kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.579996 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.580061 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.580101 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.682074 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.682144 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.682318 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrwf\" (UniqueName: \"kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.682435 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.689596 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.689680 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.695763 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.703042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrwf\" (UniqueName: \"kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:06 crc kubenswrapper[5002]: I1014 08:28:06.759375 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:07 crc kubenswrapper[5002]: I1014 08:28:07.148532 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn"] Oct 14 08:28:07 crc kubenswrapper[5002]: I1014 08:28:07.343618 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" event={"ID":"7229bab8-aa3f-46dc-bcfc-338bce4801af","Type":"ContainerStarted","Data":"b9f8db319ba2b74c9f5f58ec3969e1ed2682bb9be8b2e30f765120edfd17efb5"} Oct 14 08:28:07 crc kubenswrapper[5002]: I1014 08:28:07.697109 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:28:08 crc kubenswrapper[5002]: I1014 08:28:08.352482 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" event={"ID":"7229bab8-aa3f-46dc-bcfc-338bce4801af","Type":"ContainerStarted","Data":"0b0cab5cb1a612abaee476df8563705dc0340b8830619c2c7868114adc7d6859"} Oct 14 08:28:08 crc kubenswrapper[5002]: I1014 08:28:08.366915 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" podStartSLOduration=1.835720703 podStartE2EDuration="2.3668936s" podCreationTimestamp="2025-10-14 08:28:06 +0000 UTC" firstStartedPulling="2025-10-14 08:28:07.162767841 +0000 UTC m=+2220.144007313" lastFinishedPulling="2025-10-14 08:28:07.693940718 +0000 UTC m=+2220.675180210" observedRunningTime="2025-10-14 08:28:08.366266484 +0000 UTC m=+2221.347505946" watchObservedRunningTime="2025-10-14 08:28:08.3668936 +0000 UTC m=+2221.348133062" Oct 14 08:28:10 crc kubenswrapper[5002]: I1014 08:28:10.721388 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:28:10 crc kubenswrapper[5002]: E1014 08:28:10.723452 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:28:14 crc kubenswrapper[5002]: I1014 08:28:14.408791 5002 generic.go:334] "Generic (PLEG): container finished" podID="7229bab8-aa3f-46dc-bcfc-338bce4801af" containerID="0b0cab5cb1a612abaee476df8563705dc0340b8830619c2c7868114adc7d6859" exitCode=0 Oct 14 08:28:14 crc kubenswrapper[5002]: I1014 08:28:14.408878 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" event={"ID":"7229bab8-aa3f-46dc-bcfc-338bce4801af","Type":"ContainerDied","Data":"0b0cab5cb1a612abaee476df8563705dc0340b8830619c2c7868114adc7d6859"} Oct 14 08:28:15 crc kubenswrapper[5002]: I1014 08:28:15.951216 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.071799 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph\") pod \"7229bab8-aa3f-46dc-bcfc-338bce4801af\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.071937 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjrwf\" (UniqueName: \"kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf\") pod \"7229bab8-aa3f-46dc-bcfc-338bce4801af\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.072058 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key\") pod \"7229bab8-aa3f-46dc-bcfc-338bce4801af\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.072100 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory\") pod \"7229bab8-aa3f-46dc-bcfc-338bce4801af\" (UID: \"7229bab8-aa3f-46dc-bcfc-338bce4801af\") " Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.079418 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf" (OuterVolumeSpecName: "kube-api-access-cjrwf") pod "7229bab8-aa3f-46dc-bcfc-338bce4801af" (UID: "7229bab8-aa3f-46dc-bcfc-338bce4801af"). InnerVolumeSpecName "kube-api-access-cjrwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.084048 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph" (OuterVolumeSpecName: "ceph") pod "7229bab8-aa3f-46dc-bcfc-338bce4801af" (UID: "7229bab8-aa3f-46dc-bcfc-338bce4801af"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.110571 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory" (OuterVolumeSpecName: "inventory") pod "7229bab8-aa3f-46dc-bcfc-338bce4801af" (UID: "7229bab8-aa3f-46dc-bcfc-338bce4801af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.134199 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7229bab8-aa3f-46dc-bcfc-338bce4801af" (UID: "7229bab8-aa3f-46dc-bcfc-338bce4801af"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.174159 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.174205 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.174222 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7229bab8-aa3f-46dc-bcfc-338bce4801af-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.174240 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjrwf\" (UniqueName: \"kubernetes.io/projected/7229bab8-aa3f-46dc-bcfc-338bce4801af-kube-api-access-cjrwf\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.431774 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" event={"ID":"7229bab8-aa3f-46dc-bcfc-338bce4801af","Type":"ContainerDied","Data":"b9f8db319ba2b74c9f5f58ec3969e1ed2682bb9be8b2e30f765120edfd17efb5"} Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.432098 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9f8db319ba2b74c9f5f58ec3969e1ed2682bb9be8b2e30f765120edfd17efb5" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.431907 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.541409 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq"] Oct 14 08:28:16 crc kubenswrapper[5002]: E1014 08:28:16.541871 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7229bab8-aa3f-46dc-bcfc-338bce4801af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.541893 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7229bab8-aa3f-46dc-bcfc-338bce4801af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.542186 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7229bab8-aa3f-46dc-bcfc-338bce4801af" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.542981 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.549686 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.549907 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.556594 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.557123 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.557157 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.557341 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq"] Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.684677 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.684978 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.685051 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjj4g\" (UniqueName: \"kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.685270 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.788097 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.789152 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.789234 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjj4g\" (UniqueName: \"kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.789451 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.795041 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.795925 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.796804 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.808432 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjj4g\" (UniqueName: \"kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfkxq\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:16 crc kubenswrapper[5002]: I1014 08:28:16.876590 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:17 crc kubenswrapper[5002]: I1014 08:28:17.237461 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq"] Oct 14 08:28:17 crc kubenswrapper[5002]: I1014 08:28:17.440673 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" event={"ID":"f8e84847-a7b7-43a3-90c7-7a9c1ea92415","Type":"ContainerStarted","Data":"b76d6b1f274510373971b35d0d90f8555a9bb8781c4592d67ff81a88e2bff76f"} Oct 14 08:28:18 crc kubenswrapper[5002]: I1014 08:28:18.458312 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" event={"ID":"f8e84847-a7b7-43a3-90c7-7a9c1ea92415","Type":"ContainerStarted","Data":"33fccde0c9a7b132f8e60208bbe7244e63d193dbf2fc62b09dfd9221e5a0850d"} Oct 14 08:28:25 crc kubenswrapper[5002]: I1014 08:28:25.721052 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:28:25 crc kubenswrapper[5002]: E1014 08:28:25.721796 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:28:36 crc kubenswrapper[5002]: I1014 08:28:36.721739 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:28:36 crc kubenswrapper[5002]: E1014 08:28:36.722830 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:28:49 crc kubenswrapper[5002]: I1014 08:28:49.720706 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:28:49 crc kubenswrapper[5002]: E1014 08:28:49.722045 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:28:57 crc kubenswrapper[5002]: I1014 08:28:57.896196 5002 generic.go:334] "Generic (PLEG): container finished" podID="f8e84847-a7b7-43a3-90c7-7a9c1ea92415" containerID="33fccde0c9a7b132f8e60208bbe7244e63d193dbf2fc62b09dfd9221e5a0850d" exitCode=0 Oct 14 08:28:57 crc kubenswrapper[5002]: I1014 08:28:57.896318 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" event={"ID":"f8e84847-a7b7-43a3-90c7-7a9c1ea92415","Type":"ContainerDied","Data":"33fccde0c9a7b132f8e60208bbe7244e63d193dbf2fc62b09dfd9221e5a0850d"} Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.428084 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.539434 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key\") pod \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.539512 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjj4g\" (UniqueName: \"kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g\") pod \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.539551 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory\") pod \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.539674 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph\") pod \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\" (UID: \"f8e84847-a7b7-43a3-90c7-7a9c1ea92415\") " Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.545874 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph" (OuterVolumeSpecName: "ceph") pod "f8e84847-a7b7-43a3-90c7-7a9c1ea92415" (UID: "f8e84847-a7b7-43a3-90c7-7a9c1ea92415"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.546729 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g" (OuterVolumeSpecName: "kube-api-access-zjj4g") pod "f8e84847-a7b7-43a3-90c7-7a9c1ea92415" (UID: "f8e84847-a7b7-43a3-90c7-7a9c1ea92415"). InnerVolumeSpecName "kube-api-access-zjj4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.566307 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory" (OuterVolumeSpecName: "inventory") pod "f8e84847-a7b7-43a3-90c7-7a9c1ea92415" (UID: "f8e84847-a7b7-43a3-90c7-7a9c1ea92415"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.574691 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f8e84847-a7b7-43a3-90c7-7a9c1ea92415" (UID: "f8e84847-a7b7-43a3-90c7-7a9c1ea92415"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.641732 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.641783 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjj4g\" (UniqueName: \"kubernetes.io/projected/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-kube-api-access-zjj4g\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.641804 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.641822 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f8e84847-a7b7-43a3-90c7-7a9c1ea92415-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.922244 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" event={"ID":"f8e84847-a7b7-43a3-90c7-7a9c1ea92415","Type":"ContainerDied","Data":"b76d6b1f274510373971b35d0d90f8555a9bb8781c4592d67ff81a88e2bff76f"} Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.922293 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b76d6b1f274510373971b35d0d90f8555a9bb8781c4592d67ff81a88e2bff76f" Oct 14 08:28:59 crc kubenswrapper[5002]: I1014 08:28:59.922323 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfkxq" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.011820 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h"] Oct 14 08:29:00 crc kubenswrapper[5002]: E1014 08:29:00.012244 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8e84847-a7b7-43a3-90c7-7a9c1ea92415" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.012267 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8e84847-a7b7-43a3-90c7-7a9c1ea92415" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.012444 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8e84847-a7b7-43a3-90c7-7a9c1ea92415" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.013065 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.015121 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.015258 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.020339 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.020376 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.020384 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.031899 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h"] Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.155342 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.155458 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.155593 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.155692 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztwjk\" (UniqueName: \"kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.258698 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.258867 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.259037 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.259136 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztwjk\" (UniqueName: \"kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.266653 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.266720 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.268457 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.281318 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztwjk\" (UniqueName: \"kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.336274 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:00 crc kubenswrapper[5002]: I1014 08:29:00.925956 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h"] Oct 14 08:29:01 crc kubenswrapper[5002]: I1014 08:29:01.943973 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" event={"ID":"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec","Type":"ContainerStarted","Data":"925dc107d42cd3127e3c9529191e64624f0b8b71ff2e336163676bf31beb8739"} Oct 14 08:29:02 crc kubenswrapper[5002]: I1014 08:29:02.959725 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" event={"ID":"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec","Type":"ContainerStarted","Data":"069588b6e8f1fd063aefeaa658492322e9efa4844e3b35a7b01f2a8701f0b250"} Oct 14 08:29:03 crc kubenswrapper[5002]: I1014 08:29:03.001991 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" podStartSLOduration=3.170591649 podStartE2EDuration="4.001957213s" podCreationTimestamp="2025-10-14 08:28:59 +0000 UTC" firstStartedPulling="2025-10-14 08:29:00.93553698 +0000 UTC m=+2273.916776472" lastFinishedPulling="2025-10-14 08:29:01.766902554 +0000 UTC m=+2274.748142036" observedRunningTime="2025-10-14 08:29:02.987019851 +0000 UTC m=+2275.968259353" watchObservedRunningTime="2025-10-14 08:29:03.001957213 +0000 UTC m=+2275.983196705" Oct 14 08:29:04 crc kubenswrapper[5002]: I1014 08:29:04.720677 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:29:04 crc kubenswrapper[5002]: E1014 08:29:04.722743 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:29:07 crc kubenswrapper[5002]: I1014 08:29:07.004641 5002 generic.go:334] "Generic (PLEG): container finished" podID="b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" containerID="069588b6e8f1fd063aefeaa658492322e9efa4844e3b35a7b01f2a8701f0b250" exitCode=0 Oct 14 08:29:07 crc kubenswrapper[5002]: I1014 08:29:07.004757 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" event={"ID":"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec","Type":"ContainerDied","Data":"069588b6e8f1fd063aefeaa658492322e9efa4844e3b35a7b01f2a8701f0b250"} Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.475672 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.652369 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph\") pod \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.653833 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztwjk\" (UniqueName: \"kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk\") pod \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.654192 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key\") pod \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.654439 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory\") pod \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\" (UID: \"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec\") " Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.663817 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk" (OuterVolumeSpecName: "kube-api-access-ztwjk") pod "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" (UID: "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec"). InnerVolumeSpecName "kube-api-access-ztwjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.664550 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph" (OuterVolumeSpecName: "ceph") pod "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" (UID: "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.700204 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory" (OuterVolumeSpecName: "inventory") pod "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" (UID: "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.708026 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" (UID: "b871be8d-75a8-4c47-bfe4-4ad8f79d7fec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.758168 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.758233 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.758264 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztwjk\" (UniqueName: \"kubernetes.io/projected/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-kube-api-access-ztwjk\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:08 crc kubenswrapper[5002]: I1014 08:29:08.758291 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b871be8d-75a8-4c47-bfe4-4ad8f79d7fec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.029339 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" event={"ID":"b871be8d-75a8-4c47-bfe4-4ad8f79d7fec","Type":"ContainerDied","Data":"925dc107d42cd3127e3c9529191e64624f0b8b71ff2e336163676bf31beb8739"} Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.029397 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925dc107d42cd3127e3c9529191e64624f0b8b71ff2e336163676bf31beb8739" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.029827 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.125993 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d"] Oct 14 08:29:09 crc kubenswrapper[5002]: E1014 08:29:09.135635 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.135899 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.136515 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b871be8d-75a8-4c47-bfe4-4ad8f79d7fec" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.137148 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d"] Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.137331 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.140079 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.140174 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.141336 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.141648 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.144745 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.165397 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.165649 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7jv\" (UniqueName: \"kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.165798 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.165952 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.267470 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.267577 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.267762 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.267809 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7jv\" (UniqueName: \"kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.273576 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.274058 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.274193 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.287134 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7jv\" (UniqueName: \"kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.457076 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:29:09 crc kubenswrapper[5002]: I1014 08:29:09.775108 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d"] Oct 14 08:29:09 crc kubenswrapper[5002]: W1014 08:29:09.783053 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34674ceb_4ae7_48fb_84be_afaae15ca5bf.slice/crio-b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad WatchSource:0}: Error finding container b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad: Status 404 returned error can't find the container with id b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad Oct 14 08:29:10 crc kubenswrapper[5002]: I1014 08:29:10.043996 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" event={"ID":"34674ceb-4ae7-48fb-84be-afaae15ca5bf","Type":"ContainerStarted","Data":"b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad"} Oct 14 08:29:11 crc kubenswrapper[5002]: I1014 08:29:11.056783 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" event={"ID":"34674ceb-4ae7-48fb-84be-afaae15ca5bf","Type":"ContainerStarted","Data":"865bba6992fdcb156e812d5cbeb99657849f511994172d94f19943ca6c339b82"} Oct 14 08:29:11 crc kubenswrapper[5002]: I1014 08:29:11.088865 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" podStartSLOduration=1.65275939 podStartE2EDuration="2.088824844s" podCreationTimestamp="2025-10-14 08:29:09 +0000 UTC" firstStartedPulling="2025-10-14 08:29:09.784760773 +0000 UTC m=+2282.766000225" lastFinishedPulling="2025-10-14 08:29:10.220826187 +0000 UTC m=+2283.202065679" observedRunningTime="2025-10-14 08:29:11.077158281 +0000 UTC m=+2284.058397793" watchObservedRunningTime="2025-10-14 08:29:11.088824844 +0000 UTC m=+2284.070064306" Oct 14 08:29:18 crc kubenswrapper[5002]: I1014 08:29:18.720309 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:29:18 crc kubenswrapper[5002]: E1014 08:29:18.720976 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:29:31 crc kubenswrapper[5002]: I1014 08:29:31.721250 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:29:31 crc kubenswrapper[5002]: E1014 08:29:31.721971 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:29:42 crc kubenswrapper[5002]: I1014 08:29:42.720443 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:29:42 crc kubenswrapper[5002]: E1014 08:29:42.721214 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:29:56 crc kubenswrapper[5002]: I1014 08:29:56.720632 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:29:56 crc kubenswrapper[5002]: E1014 08:29:56.721684 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.159432 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss"] Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.162165 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.166023 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.166679 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.174551 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss"] Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.266127 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.266185 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfzg2\" (UniqueName: \"kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.266284 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.368241 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.368346 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfzg2\" (UniqueName: \"kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.368585 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.369295 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.375683 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.389473 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfzg2\" (UniqueName: \"kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2\") pod \"collect-profiles-29340510-xrbss\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.498622 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.581536 5002 generic.go:334] "Generic (PLEG): container finished" podID="34674ceb-4ae7-48fb-84be-afaae15ca5bf" containerID="865bba6992fdcb156e812d5cbeb99657849f511994172d94f19943ca6c339b82" exitCode=0 Oct 14 08:30:00 crc kubenswrapper[5002]: I1014 08:30:00.581615 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" event={"ID":"34674ceb-4ae7-48fb-84be-afaae15ca5bf","Type":"ContainerDied","Data":"865bba6992fdcb156e812d5cbeb99657849f511994172d94f19943ca6c339b82"} Oct 14 08:30:01 crc kubenswrapper[5002]: I1014 08:30:01.003244 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss"] Oct 14 08:30:01 crc kubenswrapper[5002]: I1014 08:30:01.593669 5002 generic.go:334] "Generic (PLEG): container finished" podID="82b52361-3089-4ec1-a9aa-c29bc75a7007" containerID="f1ba0fa315f025033787adb1b32273076d78ae390483a9c0a091c87e253a91ea" exitCode=0 Oct 14 08:30:01 crc kubenswrapper[5002]: I1014 08:30:01.593755 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" event={"ID":"82b52361-3089-4ec1-a9aa-c29bc75a7007","Type":"ContainerDied","Data":"f1ba0fa315f025033787adb1b32273076d78ae390483a9c0a091c87e253a91ea"} Oct 14 08:30:01 crc kubenswrapper[5002]: I1014 08:30:01.596301 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" event={"ID":"82b52361-3089-4ec1-a9aa-c29bc75a7007","Type":"ContainerStarted","Data":"ca7900e0c9e6f709d00de9e7c497456ece5df8efa1d87bde6a323478508733ad"} Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.110924 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.202774 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory\") pod \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.203070 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv7jv\" (UniqueName: \"kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv\") pod \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.203123 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key\") pod \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.203149 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph\") pod \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\" (UID: \"34674ceb-4ae7-48fb-84be-afaae15ca5bf\") " Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.212056 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv" (OuterVolumeSpecName: "kube-api-access-vv7jv") pod "34674ceb-4ae7-48fb-84be-afaae15ca5bf" (UID: "34674ceb-4ae7-48fb-84be-afaae15ca5bf"). InnerVolumeSpecName "kube-api-access-vv7jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.212077 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph" (OuterVolumeSpecName: "ceph") pod "34674ceb-4ae7-48fb-84be-afaae15ca5bf" (UID: "34674ceb-4ae7-48fb-84be-afaae15ca5bf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.238341 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory" (OuterVolumeSpecName: "inventory") pod "34674ceb-4ae7-48fb-84be-afaae15ca5bf" (UID: "34674ceb-4ae7-48fb-84be-afaae15ca5bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.239721 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34674ceb-4ae7-48fb-84be-afaae15ca5bf" (UID: "34674ceb-4ae7-48fb-84be-afaae15ca5bf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.305754 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv7jv\" (UniqueName: \"kubernetes.io/projected/34674ceb-4ae7-48fb-84be-afaae15ca5bf-kube-api-access-vv7jv\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.305805 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.305818 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.305855 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34674ceb-4ae7-48fb-84be-afaae15ca5bf-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.612200 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.612340 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d" event={"ID":"34674ceb-4ae7-48fb-84be-afaae15ca5bf","Type":"ContainerDied","Data":"b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad"} Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.612760 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b67ea82a6b0728d24184ca3183c44f26cdd55e0c1398cd1b6fcb86707f249bad" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.712824 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-46bxg"] Oct 14 08:30:02 crc kubenswrapper[5002]: E1014 08:30:02.713191 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34674ceb-4ae7-48fb-84be-afaae15ca5bf" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.713203 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="34674ceb-4ae7-48fb-84be-afaae15ca5bf" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.713384 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="34674ceb-4ae7-48fb-84be-afaae15ca5bf" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.714010 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.717943 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-46bxg"] Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.720380 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.720380 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.720436 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.720762 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.721270 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.818470 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzld7\" (UniqueName: \"kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.818686 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.818735 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.819024 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.920155 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzld7\" (UniqueName: \"kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.920279 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.920302 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.920346 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.925497 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.925784 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.926030 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:02 crc kubenswrapper[5002]: I1014 08:30:02.954671 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzld7\" (UniqueName: \"kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7\") pod \"ssh-known-hosts-edpm-deployment-46bxg\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.002599 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.045988 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.122810 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume\") pod \"82b52361-3089-4ec1-a9aa-c29bc75a7007\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.123174 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume\") pod \"82b52361-3089-4ec1-a9aa-c29bc75a7007\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.123208 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfzg2\" (UniqueName: \"kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2\") pod \"82b52361-3089-4ec1-a9aa-c29bc75a7007\" (UID: \"82b52361-3089-4ec1-a9aa-c29bc75a7007\") " Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.123788 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume" (OuterVolumeSpecName: "config-volume") pod "82b52361-3089-4ec1-a9aa-c29bc75a7007" (UID: "82b52361-3089-4ec1-a9aa-c29bc75a7007"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.127100 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2" (OuterVolumeSpecName: "kube-api-access-lfzg2") pod "82b52361-3089-4ec1-a9aa-c29bc75a7007" (UID: "82b52361-3089-4ec1-a9aa-c29bc75a7007"). InnerVolumeSpecName "kube-api-access-lfzg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.127407 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "82b52361-3089-4ec1-a9aa-c29bc75a7007" (UID: "82b52361-3089-4ec1-a9aa-c29bc75a7007"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.225232 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/82b52361-3089-4ec1-a9aa-c29bc75a7007-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.225262 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/82b52361-3089-4ec1-a9aa-c29bc75a7007-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.225273 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfzg2\" (UniqueName: \"kubernetes.io/projected/82b52361-3089-4ec1-a9aa-c29bc75a7007-kube-api-access-lfzg2\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.565699 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-46bxg"] Oct 14 08:30:03 crc kubenswrapper[5002]: W1014 08:30:03.571368 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1704867_21c5_46a1_8d37_85af1337b303.slice/crio-a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b WatchSource:0}: Error finding container a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b: Status 404 returned error can't find the container with id a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.621155 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" event={"ID":"82b52361-3089-4ec1-a9aa-c29bc75a7007","Type":"ContainerDied","Data":"ca7900e0c9e6f709d00de9e7c497456ece5df8efa1d87bde6a323478508733ad"} Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.621184 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.621208 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca7900e0c9e6f709d00de9e7c497456ece5df8efa1d87bde6a323478508733ad" Oct 14 08:30:03 crc kubenswrapper[5002]: I1014 08:30:03.622233 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" event={"ID":"b1704867-21c5-46a1-8d37-85af1337b303","Type":"ContainerStarted","Data":"a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b"} Oct 14 08:30:04 crc kubenswrapper[5002]: I1014 08:30:04.100268 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm"] Oct 14 08:30:04 crc kubenswrapper[5002]: I1014 08:30:04.108275 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-h7nfm"] Oct 14 08:30:05 crc kubenswrapper[5002]: I1014 08:30:05.648485 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" event={"ID":"b1704867-21c5-46a1-8d37-85af1337b303","Type":"ContainerStarted","Data":"2e179ec55f0f7756519dffc398a0dc94c181fbf0a9c225f86e08d367352394c6"} Oct 14 08:30:05 crc kubenswrapper[5002]: I1014 08:30:05.685710 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" podStartSLOduration=2.811271111 podStartE2EDuration="3.6856822s" podCreationTimestamp="2025-10-14 08:30:02 +0000 UTC" firstStartedPulling="2025-10-14 08:30:03.573745445 +0000 UTC m=+2336.554984907" lastFinishedPulling="2025-10-14 08:30:04.448156504 +0000 UTC m=+2337.429395996" observedRunningTime="2025-10-14 08:30:05.673092082 +0000 UTC m=+2338.654331594" watchObservedRunningTime="2025-10-14 08:30:05.6856822 +0000 UTC m=+2338.666921692" Oct 14 08:30:05 crc kubenswrapper[5002]: I1014 08:30:05.738365 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f63d16a-af72-4d47-828d-8ac2532a6196" path="/var/lib/kubelet/pods/8f63d16a-af72-4d47-828d-8ac2532a6196/volumes" Oct 14 08:30:08 crc kubenswrapper[5002]: I1014 08:30:08.721687 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:30:08 crc kubenswrapper[5002]: E1014 08:30:08.722730 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:30:15 crc kubenswrapper[5002]: I1014 08:30:15.755049 5002 generic.go:334] "Generic (PLEG): container finished" podID="b1704867-21c5-46a1-8d37-85af1337b303" containerID="2e179ec55f0f7756519dffc398a0dc94c181fbf0a9c225f86e08d367352394c6" exitCode=0 Oct 14 08:30:15 crc kubenswrapper[5002]: I1014 08:30:15.755198 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" event={"ID":"b1704867-21c5-46a1-8d37-85af1337b303","Type":"ContainerDied","Data":"2e179ec55f0f7756519dffc398a0dc94c181fbf0a9c225f86e08d367352394c6"} Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.289197 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.330966 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzld7\" (UniqueName: \"kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7\") pod \"b1704867-21c5-46a1-8d37-85af1337b303\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.331137 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph\") pod \"b1704867-21c5-46a1-8d37-85af1337b303\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.331181 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0\") pod \"b1704867-21c5-46a1-8d37-85af1337b303\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.331326 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam\") pod \"b1704867-21c5-46a1-8d37-85af1337b303\" (UID: \"b1704867-21c5-46a1-8d37-85af1337b303\") " Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.373359 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7" (OuterVolumeSpecName: "kube-api-access-bzld7") pod "b1704867-21c5-46a1-8d37-85af1337b303" (UID: "b1704867-21c5-46a1-8d37-85af1337b303"). InnerVolumeSpecName "kube-api-access-bzld7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.373981 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph" (OuterVolumeSpecName: "ceph") pod "b1704867-21c5-46a1-8d37-85af1337b303" (UID: "b1704867-21c5-46a1-8d37-85af1337b303"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.432432 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzld7\" (UniqueName: \"kubernetes.io/projected/b1704867-21c5-46a1-8d37-85af1337b303-kube-api-access-bzld7\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.432467 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.432626 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b1704867-21c5-46a1-8d37-85af1337b303" (UID: "b1704867-21c5-46a1-8d37-85af1337b303"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.442049 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b1704867-21c5-46a1-8d37-85af1337b303" (UID: "b1704867-21c5-46a1-8d37-85af1337b303"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.533913 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.533956 5002 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b1704867-21c5-46a1-8d37-85af1337b303-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.778649 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" event={"ID":"b1704867-21c5-46a1-8d37-85af1337b303","Type":"ContainerDied","Data":"a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b"} Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.779464 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9f602d6f339f6863990e999e47a7288ed9d760009ad1b031eaa555dd6fe6b5b" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.778733 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-46bxg" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.868038 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l"] Oct 14 08:30:17 crc kubenswrapper[5002]: E1014 08:30:17.868568 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1704867-21c5-46a1-8d37-85af1337b303" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.868599 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1704867-21c5-46a1-8d37-85af1337b303" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:30:17 crc kubenswrapper[5002]: E1014 08:30:17.868636 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b52361-3089-4ec1-a9aa-c29bc75a7007" containerName="collect-profiles" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.868650 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b52361-3089-4ec1-a9aa-c29bc75a7007" containerName="collect-profiles" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.869033 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b52361-3089-4ec1-a9aa-c29bc75a7007" containerName="collect-profiles" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.869088 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1704867-21c5-46a1-8d37-85af1337b303" containerName="ssh-known-hosts-edpm-deployment" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.870131 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.876319 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.876549 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.876587 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.876777 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.877206 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:30:17 crc kubenswrapper[5002]: I1014 08:30:17.880153 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l"] Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.044855 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.044945 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr76q\" (UniqueName: \"kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.045297 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.045401 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.147049 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.147152 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.147324 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.147405 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr76q\" (UniqueName: \"kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.153091 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.153789 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.153987 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.184994 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr76q\" (UniqueName: \"kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-hlm2l\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.196035 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.768147 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l"] Oct 14 08:30:18 crc kubenswrapper[5002]: I1014 08:30:18.787891 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" event={"ID":"75c3bce2-2453-4d20-a946-9a28fe2151e8","Type":"ContainerStarted","Data":"3719600181a7a4d8655db788987f41af82c316aef949c463eebb07eecd9217b7"} Oct 14 08:30:19 crc kubenswrapper[5002]: I1014 08:30:19.799339 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" event={"ID":"75c3bce2-2453-4d20-a946-9a28fe2151e8","Type":"ContainerStarted","Data":"9a5bcba6bd83b760eb2fb57db59aa85ad140ecd2ca2077d1feea45ad2a778996"} Oct 14 08:30:19 crc kubenswrapper[5002]: I1014 08:30:19.826204 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" podStartSLOduration=2.221553263 podStartE2EDuration="2.826175131s" podCreationTimestamp="2025-10-14 08:30:17 +0000 UTC" firstStartedPulling="2025-10-14 08:30:18.7721378 +0000 UTC m=+2351.753377252" lastFinishedPulling="2025-10-14 08:30:19.376759648 +0000 UTC m=+2352.357999120" observedRunningTime="2025-10-14 08:30:19.81906462 +0000 UTC m=+2352.800304092" watchObservedRunningTime="2025-10-14 08:30:19.826175131 +0000 UTC m=+2352.807414583" Oct 14 08:30:20 crc kubenswrapper[5002]: I1014 08:30:20.721119 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:30:20 crc kubenswrapper[5002]: E1014 08:30:20.721477 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:30:25 crc kubenswrapper[5002]: I1014 08:30:25.086664 5002 scope.go:117] "RemoveContainer" containerID="7c39131f69a4475e22d2bbdfba61f63c57ae6c93d2ec3e1ff522c9724b63aa2d" Oct 14 08:30:27 crc kubenswrapper[5002]: I1014 08:30:27.876357 5002 generic.go:334] "Generic (PLEG): container finished" podID="75c3bce2-2453-4d20-a946-9a28fe2151e8" containerID="9a5bcba6bd83b760eb2fb57db59aa85ad140ecd2ca2077d1feea45ad2a778996" exitCode=0 Oct 14 08:30:27 crc kubenswrapper[5002]: I1014 08:30:27.876422 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" event={"ID":"75c3bce2-2453-4d20-a946-9a28fe2151e8","Type":"ContainerDied","Data":"9a5bcba6bd83b760eb2fb57db59aa85ad140ecd2ca2077d1feea45ad2a778996"} Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.388983 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.549929 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr76q\" (UniqueName: \"kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q\") pod \"75c3bce2-2453-4d20-a946-9a28fe2151e8\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.549979 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory\") pod \"75c3bce2-2453-4d20-a946-9a28fe2151e8\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.550209 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key\") pod \"75c3bce2-2453-4d20-a946-9a28fe2151e8\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.550261 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph\") pod \"75c3bce2-2453-4d20-a946-9a28fe2151e8\" (UID: \"75c3bce2-2453-4d20-a946-9a28fe2151e8\") " Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.555930 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph" (OuterVolumeSpecName: "ceph") pod "75c3bce2-2453-4d20-a946-9a28fe2151e8" (UID: "75c3bce2-2453-4d20-a946-9a28fe2151e8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.557041 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q" (OuterVolumeSpecName: "kube-api-access-pr76q") pod "75c3bce2-2453-4d20-a946-9a28fe2151e8" (UID: "75c3bce2-2453-4d20-a946-9a28fe2151e8"). InnerVolumeSpecName "kube-api-access-pr76q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.574270 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "75c3bce2-2453-4d20-a946-9a28fe2151e8" (UID: "75c3bce2-2453-4d20-a946-9a28fe2151e8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.574447 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory" (OuterVolumeSpecName: "inventory") pod "75c3bce2-2453-4d20-a946-9a28fe2151e8" (UID: "75c3bce2-2453-4d20-a946-9a28fe2151e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.651865 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.651903 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.651915 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr76q\" (UniqueName: \"kubernetes.io/projected/75c3bce2-2453-4d20-a946-9a28fe2151e8-kube-api-access-pr76q\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.651930 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75c3bce2-2453-4d20-a946-9a28fe2151e8-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.898187 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" event={"ID":"75c3bce2-2453-4d20-a946-9a28fe2151e8","Type":"ContainerDied","Data":"3719600181a7a4d8655db788987f41af82c316aef949c463eebb07eecd9217b7"} Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.898499 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3719600181a7a4d8655db788987f41af82c316aef949c463eebb07eecd9217b7" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.898254 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-hlm2l" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.987271 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z"] Oct 14 08:30:29 crc kubenswrapper[5002]: E1014 08:30:29.987747 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75c3bce2-2453-4d20-a946-9a28fe2151e8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.987772 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="75c3bce2-2453-4d20-a946-9a28fe2151e8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.988014 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="75c3bce2-2453-4d20-a946-9a28fe2151e8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.988797 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.993309 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.993612 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.993627 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.993806 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:30:29 crc kubenswrapper[5002]: I1014 08:30:29.994175 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.000210 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z"] Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.160517 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.160602 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.160709 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfjqc\" (UniqueName: \"kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.160757 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.263256 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.263443 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.263524 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfjqc\" (UniqueName: \"kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.263618 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.270146 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.270526 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.280582 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.298555 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfjqc\" (UniqueName: \"kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-f955z\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.307074 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.835201 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z"] Oct 14 08:30:30 crc kubenswrapper[5002]: I1014 08:30:30.910517 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" event={"ID":"3b28c2e3-928a-400a-92e3-c4e407257f94","Type":"ContainerStarted","Data":"e6daddf1f6dc5033f4529ee3376ef3628efca08e16368e9cbfddfc65ca4bd948"} Oct 14 08:30:32 crc kubenswrapper[5002]: I1014 08:30:32.933706 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" event={"ID":"3b28c2e3-928a-400a-92e3-c4e407257f94","Type":"ContainerStarted","Data":"a6a9ddf0a3d736b4a07254482ae2aab498a5919b4067012c3cd745d8be599cad"} Oct 14 08:30:32 crc kubenswrapper[5002]: I1014 08:30:32.958413 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" podStartSLOduration=3.105025266 podStartE2EDuration="3.95838777s" podCreationTimestamp="2025-10-14 08:30:29 +0000 UTC" firstStartedPulling="2025-10-14 08:30:30.844048741 +0000 UTC m=+2363.825288213" lastFinishedPulling="2025-10-14 08:30:31.697411225 +0000 UTC m=+2364.678650717" observedRunningTime="2025-10-14 08:30:32.955556794 +0000 UTC m=+2365.936796316" watchObservedRunningTime="2025-10-14 08:30:32.95838777 +0000 UTC m=+2365.939627232" Oct 14 08:30:33 crc kubenswrapper[5002]: I1014 08:30:33.720623 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:30:33 crc kubenswrapper[5002]: E1014 08:30:33.721385 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:30:43 crc kubenswrapper[5002]: I1014 08:30:43.037576 5002 generic.go:334] "Generic (PLEG): container finished" podID="3b28c2e3-928a-400a-92e3-c4e407257f94" containerID="a6a9ddf0a3d736b4a07254482ae2aab498a5919b4067012c3cd745d8be599cad" exitCode=0 Oct 14 08:30:43 crc kubenswrapper[5002]: I1014 08:30:43.037663 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" event={"ID":"3b28c2e3-928a-400a-92e3-c4e407257f94","Type":"ContainerDied","Data":"a6a9ddf0a3d736b4a07254482ae2aab498a5919b4067012c3cd745d8be599cad"} Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.579036 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.755274 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key\") pod \"3b28c2e3-928a-400a-92e3-c4e407257f94\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.755519 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfjqc\" (UniqueName: \"kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc\") pod \"3b28c2e3-928a-400a-92e3-c4e407257f94\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.755606 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph\") pod \"3b28c2e3-928a-400a-92e3-c4e407257f94\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.755868 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory\") pod \"3b28c2e3-928a-400a-92e3-c4e407257f94\" (UID: \"3b28c2e3-928a-400a-92e3-c4e407257f94\") " Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.764652 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc" (OuterVolumeSpecName: "kube-api-access-nfjqc") pod "3b28c2e3-928a-400a-92e3-c4e407257f94" (UID: "3b28c2e3-928a-400a-92e3-c4e407257f94"). InnerVolumeSpecName "kube-api-access-nfjqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.765572 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph" (OuterVolumeSpecName: "ceph") pod "3b28c2e3-928a-400a-92e3-c4e407257f94" (UID: "3b28c2e3-928a-400a-92e3-c4e407257f94"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.804142 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b28c2e3-928a-400a-92e3-c4e407257f94" (UID: "3b28c2e3-928a-400a-92e3-c4e407257f94"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.805527 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory" (OuterVolumeSpecName: "inventory") pod "3b28c2e3-928a-400a-92e3-c4e407257f94" (UID: "3b28c2e3-928a-400a-92e3-c4e407257f94"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.859785 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.859828 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfjqc\" (UniqueName: \"kubernetes.io/projected/3b28c2e3-928a-400a-92e3-c4e407257f94-kube-api-access-nfjqc\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.859875 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:44 crc kubenswrapper[5002]: I1014 08:30:44.859894 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b28c2e3-928a-400a-92e3-c4e407257f94-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.073032 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" event={"ID":"3b28c2e3-928a-400a-92e3-c4e407257f94","Type":"ContainerDied","Data":"e6daddf1f6dc5033f4529ee3376ef3628efca08e16368e9cbfddfc65ca4bd948"} Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.073075 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6daddf1f6dc5033f4529ee3376ef3628efca08e16368e9cbfddfc65ca4bd948" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.073138 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-f955z" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.260718 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf"] Oct 14 08:30:45 crc kubenswrapper[5002]: E1014 08:30:45.261089 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b28c2e3-928a-400a-92e3-c4e407257f94" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.261106 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b28c2e3-928a-400a-92e3-c4e407257f94" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.261268 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b28c2e3-928a-400a-92e3-c4e407257f94" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.261852 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.263960 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.264203 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.266662 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.267039 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.267210 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.267366 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.267554 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.267784 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275137 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275225 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275355 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275417 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8l9v\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275448 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275472 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275529 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275601 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275637 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275675 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275692 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275739 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.275760 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.291784 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf"] Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.377799 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.377872 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.377919 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.377956 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.377984 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378009 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378025 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378053 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378071 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378103 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378133 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378211 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.378235 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8l9v\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.382544 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.383789 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.384022 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.384089 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.384161 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.384422 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.384947 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.386133 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.386359 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.386894 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.387034 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.391667 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.396878 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8l9v\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.585779 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.722033 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:30:45 crc kubenswrapper[5002]: E1014 08:30:45.722408 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.981763 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf"] Oct 14 08:30:45 crc kubenswrapper[5002]: W1014 08:30:45.985001 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bbe90c4_4bf2_4f7e_ae65_8ac5a5324819.slice/crio-0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191 WatchSource:0}: Error finding container 0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191: Status 404 returned error can't find the container with id 0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191 Oct 14 08:30:45 crc kubenswrapper[5002]: I1014 08:30:45.989382 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:30:46 crc kubenswrapper[5002]: I1014 08:30:46.083794 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" event={"ID":"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819","Type":"ContainerStarted","Data":"0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191"} Oct 14 08:30:47 crc kubenswrapper[5002]: I1014 08:30:47.095952 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" event={"ID":"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819","Type":"ContainerStarted","Data":"82ce15c3b90bb9a0a2a1064f37eadad8a1f5e579f828bed6158c7472656417b6"} Oct 14 08:30:47 crc kubenswrapper[5002]: I1014 08:30:47.130081 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" podStartSLOduration=1.403137189 podStartE2EDuration="2.130016379s" podCreationTimestamp="2025-10-14 08:30:45 +0000 UTC" firstStartedPulling="2025-10-14 08:30:45.989068776 +0000 UTC m=+2378.970308248" lastFinishedPulling="2025-10-14 08:30:46.715947946 +0000 UTC m=+2379.697187438" observedRunningTime="2025-10-14 08:30:47.125087677 +0000 UTC m=+2380.106327169" watchObservedRunningTime="2025-10-14 08:30:47.130016379 +0000 UTC m=+2380.111255861" Oct 14 08:31:00 crc kubenswrapper[5002]: I1014 08:31:00.720019 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:31:00 crc kubenswrapper[5002]: E1014 08:31:00.720792 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:31:14 crc kubenswrapper[5002]: I1014 08:31:14.721046 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:31:14 crc kubenswrapper[5002]: E1014 08:31:14.722196 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:31:25 crc kubenswrapper[5002]: I1014 08:31:25.520894 5002 generic.go:334] "Generic (PLEG): container finished" podID="1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" containerID="82ce15c3b90bb9a0a2a1064f37eadad8a1f5e579f828bed6158c7472656417b6" exitCode=0 Oct 14 08:31:25 crc kubenswrapper[5002]: I1014 08:31:25.520986 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" event={"ID":"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819","Type":"ContainerDied","Data":"82ce15c3b90bb9a0a2a1064f37eadad8a1f5e579f828bed6158c7472656417b6"} Oct 14 08:31:25 crc kubenswrapper[5002]: I1014 08:31:25.721387 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:31:25 crc kubenswrapper[5002]: E1014 08:31:25.722006 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:31:26 crc kubenswrapper[5002]: I1014 08:31:26.946076 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.087741 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8l9v\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.087794 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.087906 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.087932 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.087984 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088031 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088139 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088174 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088206 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088230 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088297 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088344 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.088368 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\" (UID: \"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819\") " Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.095168 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.095226 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.095686 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.096244 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph" (OuterVolumeSpecName: "ceph") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.096513 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v" (OuterVolumeSpecName: "kube-api-access-g8l9v") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "kube-api-access-g8l9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.097031 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.097067 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.098820 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.099195 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.100079 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.102236 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.124451 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.143260 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory" (OuterVolumeSpecName: "inventory") pod "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" (UID: "1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.190935 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.190997 5002 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191024 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191047 5002 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191069 5002 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191090 5002 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191112 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191133 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191152 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191174 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8l9v\" (UniqueName: \"kubernetes.io/projected/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-kube-api-access-g8l9v\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191192 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191210 5002 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.191230 5002 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.543006 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.543142 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf" event={"ID":"1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819","Type":"ContainerDied","Data":"0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191"} Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.547173 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d47aa19b04dc50de7243589db234f81aa037442a00bd1a60263fc035a5d5191" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.711788 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r"] Oct 14 08:31:27 crc kubenswrapper[5002]: E1014 08:31:27.712158 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.712178 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.712398 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.713082 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.715240 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.715682 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.716039 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.716704 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.723007 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.744767 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r"] Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.905508 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.905585 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.905635 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:27 crc kubenswrapper[5002]: I1014 08:31:27.905827 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnhk\" (UniqueName: \"kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.013986 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnhk\" (UniqueName: \"kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.014126 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.014163 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.014203 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.019758 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.020812 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.022256 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.032047 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnhk\" (UniqueName: \"kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.329426 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:28 crc kubenswrapper[5002]: I1014 08:31:28.909557 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r"] Oct 14 08:31:29 crc kubenswrapper[5002]: I1014 08:31:29.573849 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" event={"ID":"d93bac7d-9b37-4cfe-98dd-271aeff235ff","Type":"ContainerStarted","Data":"4edf17ebf40f089ceef140fbe6f4ba76e68e1afaaa5693ae8e659f33b9b047d4"} Oct 14 08:31:30 crc kubenswrapper[5002]: I1014 08:31:30.592882 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" event={"ID":"d93bac7d-9b37-4cfe-98dd-271aeff235ff","Type":"ContainerStarted","Data":"8b56bc5112d88554fc78aa898bf73f3fe3336876d944f033961572996d339b29"} Oct 14 08:31:30 crc kubenswrapper[5002]: I1014 08:31:30.629750 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" podStartSLOduration=2.870528957 podStartE2EDuration="3.629721528s" podCreationTimestamp="2025-10-14 08:31:27 +0000 UTC" firstStartedPulling="2025-10-14 08:31:28.913894342 +0000 UTC m=+2421.895133834" lastFinishedPulling="2025-10-14 08:31:29.673086933 +0000 UTC m=+2422.654326405" observedRunningTime="2025-10-14 08:31:30.622098274 +0000 UTC m=+2423.603337766" watchObservedRunningTime="2025-10-14 08:31:30.629721528 +0000 UTC m=+2423.610961020" Oct 14 08:31:36 crc kubenswrapper[5002]: I1014 08:31:36.660524 5002 generic.go:334] "Generic (PLEG): container finished" podID="d93bac7d-9b37-4cfe-98dd-271aeff235ff" containerID="8b56bc5112d88554fc78aa898bf73f3fe3336876d944f033961572996d339b29" exitCode=0 Oct 14 08:31:36 crc kubenswrapper[5002]: I1014 08:31:36.660595 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" event={"ID":"d93bac7d-9b37-4cfe-98dd-271aeff235ff","Type":"ContainerDied","Data":"8b56bc5112d88554fc78aa898bf73f3fe3336876d944f033961572996d339b29"} Oct 14 08:31:37 crc kubenswrapper[5002]: I1014 08:31:37.733795 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:31:37 crc kubenswrapper[5002]: E1014 08:31:37.735066 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.187656 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.261048 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key\") pod \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.261183 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph\") pod \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.261241 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory\") pod \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.297597 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph" (OuterVolumeSpecName: "ceph") pod "d93bac7d-9b37-4cfe-98dd-271aeff235ff" (UID: "d93bac7d-9b37-4cfe-98dd-271aeff235ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.300462 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory" (OuterVolumeSpecName: "inventory") pod "d93bac7d-9b37-4cfe-98dd-271aeff235ff" (UID: "d93bac7d-9b37-4cfe-98dd-271aeff235ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.301202 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d93bac7d-9b37-4cfe-98dd-271aeff235ff" (UID: "d93bac7d-9b37-4cfe-98dd-271aeff235ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.363346 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttnhk\" (UniqueName: \"kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk\") pod \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\" (UID: \"d93bac7d-9b37-4cfe-98dd-271aeff235ff\") " Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.363971 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.363994 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.364008 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93bac7d-9b37-4cfe-98dd-271aeff235ff-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.366668 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk" (OuterVolumeSpecName: "kube-api-access-ttnhk") pod "d93bac7d-9b37-4cfe-98dd-271aeff235ff" (UID: "d93bac7d-9b37-4cfe-98dd-271aeff235ff"). InnerVolumeSpecName "kube-api-access-ttnhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.464616 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttnhk\" (UniqueName: \"kubernetes.io/projected/d93bac7d-9b37-4cfe-98dd-271aeff235ff-kube-api-access-ttnhk\") on node \"crc\" DevicePath \"\"" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.685095 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" event={"ID":"d93bac7d-9b37-4cfe-98dd-271aeff235ff","Type":"ContainerDied","Data":"4edf17ebf40f089ceef140fbe6f4ba76e68e1afaaa5693ae8e659f33b9b047d4"} Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.685163 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4edf17ebf40f089ceef140fbe6f4ba76e68e1afaaa5693ae8e659f33b9b047d4" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.685175 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.793433 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g"] Oct 14 08:31:38 crc kubenswrapper[5002]: E1014 08:31:38.793872 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93bac7d-9b37-4cfe-98dd-271aeff235ff" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.793889 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93bac7d-9b37-4cfe-98dd-271aeff235ff" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.794121 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93bac7d-9b37-4cfe-98dd-271aeff235ff" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.794861 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.798562 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.798698 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.798762 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.799042 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.799967 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.806360 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.819885 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g"] Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.872588 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.872672 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.872719 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.872907 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.873181 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9sg7\" (UniqueName: \"kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.873351 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.975236 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.975490 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9sg7\" (UniqueName: \"kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.975696 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.975905 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.976034 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.976085 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.976656 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.980534 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.980954 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.981807 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:38 crc kubenswrapper[5002]: I1014 08:31:38.981830 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:39 crc kubenswrapper[5002]: I1014 08:31:39.009357 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9sg7\" (UniqueName: \"kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-sz55g\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:39 crc kubenswrapper[5002]: I1014 08:31:39.117071 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:31:39 crc kubenswrapper[5002]: I1014 08:31:39.497535 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g"] Oct 14 08:31:39 crc kubenswrapper[5002]: I1014 08:31:39.697067 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" event={"ID":"0b20c844-7dba-4cdf-b6a2-b7deff3e401a","Type":"ContainerStarted","Data":"44a93b6befe136a6ca0441a225fc3227f66610b41568ce0b3b25c866079dd1a7"} Oct 14 08:31:40 crc kubenswrapper[5002]: I1014 08:31:40.710903 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" event={"ID":"0b20c844-7dba-4cdf-b6a2-b7deff3e401a","Type":"ContainerStarted","Data":"4ac014a2a0606d96e80f2ad47c355dc9dda9cb41a6ccebb7f7f1a669f72beda8"} Oct 14 08:31:40 crc kubenswrapper[5002]: I1014 08:31:40.736513 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" podStartSLOduration=2.02936789 podStartE2EDuration="2.736496034s" podCreationTimestamp="2025-10-14 08:31:38 +0000 UTC" firstStartedPulling="2025-10-14 08:31:39.509446295 +0000 UTC m=+2432.490685747" lastFinishedPulling="2025-10-14 08:31:40.216574439 +0000 UTC m=+2433.197813891" observedRunningTime="2025-10-14 08:31:40.735961911 +0000 UTC m=+2433.717201383" watchObservedRunningTime="2025-10-14 08:31:40.736496034 +0000 UTC m=+2433.717735486" Oct 14 08:31:51 crc kubenswrapper[5002]: I1014 08:31:51.721190 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:31:51 crc kubenswrapper[5002]: E1014 08:31:51.724684 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:32:04 crc kubenswrapper[5002]: I1014 08:32:04.720961 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:32:04 crc kubenswrapper[5002]: E1014 08:32:04.722193 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:32:17 crc kubenswrapper[5002]: I1014 08:32:17.732042 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:32:18 crc kubenswrapper[5002]: I1014 08:32:18.096421 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7"} Oct 14 08:32:58 crc kubenswrapper[5002]: I1014 08:32:58.534585 5002 generic.go:334] "Generic (PLEG): container finished" podID="0b20c844-7dba-4cdf-b6a2-b7deff3e401a" containerID="4ac014a2a0606d96e80f2ad47c355dc9dda9cb41a6ccebb7f7f1a669f72beda8" exitCode=0 Oct 14 08:32:58 crc kubenswrapper[5002]: I1014 08:32:58.534633 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" event={"ID":"0b20c844-7dba-4cdf-b6a2-b7deff3e401a","Type":"ContainerDied","Data":"4ac014a2a0606d96e80f2ad47c355dc9dda9cb41a6ccebb7f7f1a669f72beda8"} Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.064770 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248160 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248286 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248313 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248341 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248465 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.248521 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9sg7\" (UniqueName: \"kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7\") pod \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\" (UID: \"0b20c844-7dba-4cdf-b6a2-b7deff3e401a\") " Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.253504 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph" (OuterVolumeSpecName: "ceph") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.254293 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.255041 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7" (OuterVolumeSpecName: "kube-api-access-g9sg7") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "kube-api-access-g9sg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.274264 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.286754 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory" (OuterVolumeSpecName: "inventory") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.287302 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b20c844-7dba-4cdf-b6a2-b7deff3e401a" (UID: "0b20c844-7dba-4cdf-b6a2-b7deff3e401a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350638 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350666 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350676 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350687 5002 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350699 5002 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.350708 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9sg7\" (UniqueName: \"kubernetes.io/projected/0b20c844-7dba-4cdf-b6a2-b7deff3e401a-kube-api-access-g9sg7\") on node \"crc\" DevicePath \"\"" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.554605 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" event={"ID":"0b20c844-7dba-4cdf-b6a2-b7deff3e401a","Type":"ContainerDied","Data":"44a93b6befe136a6ca0441a225fc3227f66610b41568ce0b3b25c866079dd1a7"} Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.554647 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-sz55g" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.554663 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a93b6befe136a6ca0441a225fc3227f66610b41568ce0b3b25c866079dd1a7" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.717009 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b"] Oct 14 08:33:00 crc kubenswrapper[5002]: E1014 08:33:00.717363 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b20c844-7dba-4cdf-b6a2-b7deff3e401a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.717383 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b20c844-7dba-4cdf-b6a2-b7deff3e401a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.717560 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b20c844-7dba-4cdf-b6a2-b7deff3e401a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.718149 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.721657 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.724429 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.724735 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.724850 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.724858 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.726003 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.726178 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.738226 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b"] Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758519 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758584 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758625 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758798 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffmff\" (UniqueName: \"kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758934 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.758990 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.759070 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861063 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861132 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861162 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861232 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffmff\" (UniqueName: \"kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861328 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.861848 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.862001 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.865639 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.866596 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.867972 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.868142 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.868263 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.881312 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:00 crc kubenswrapper[5002]: I1014 08:33:00.886807 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffmff\" (UniqueName: \"kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:01 crc kubenswrapper[5002]: I1014 08:33:01.041917 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:33:01 crc kubenswrapper[5002]: I1014 08:33:01.612812 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b"] Oct 14 08:33:02 crc kubenswrapper[5002]: I1014 08:33:02.575779 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" event={"ID":"721fe1de-3687-4a8c-babb-e30f9ec46abe","Type":"ContainerStarted","Data":"ffbf7133789dcbb11277841af0e98bb8f74d2c85159119440659277e9421170c"} Oct 14 08:33:02 crc kubenswrapper[5002]: I1014 08:33:02.576424 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" event={"ID":"721fe1de-3687-4a8c-babb-e30f9ec46abe","Type":"ContainerStarted","Data":"5bab58fc20c69c3a66e174349db696b3ca284945cb9e60137a457ab9cc85c0a6"} Oct 14 08:33:02 crc kubenswrapper[5002]: I1014 08:33:02.595821 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" podStartSLOduration=1.895672395 podStartE2EDuration="2.595801202s" podCreationTimestamp="2025-10-14 08:33:00 +0000 UTC" firstStartedPulling="2025-10-14 08:33:01.619566607 +0000 UTC m=+2514.600806059" lastFinishedPulling="2025-10-14 08:33:02.319695394 +0000 UTC m=+2515.300934866" observedRunningTime="2025-10-14 08:33:02.593875021 +0000 UTC m=+2515.575114463" watchObservedRunningTime="2025-10-14 08:33:02.595801202 +0000 UTC m=+2515.577040664" Oct 14 08:34:06 crc kubenswrapper[5002]: I1014 08:34:06.256603 5002 generic.go:334] "Generic (PLEG): container finished" podID="721fe1de-3687-4a8c-babb-e30f9ec46abe" containerID="ffbf7133789dcbb11277841af0e98bb8f74d2c85159119440659277e9421170c" exitCode=0 Oct 14 08:34:06 crc kubenswrapper[5002]: I1014 08:34:06.256674 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" event={"ID":"721fe1de-3687-4a8c-babb-e30f9ec46abe","Type":"ContainerDied","Data":"ffbf7133789dcbb11277841af0e98bb8f74d2c85159119440659277e9421170c"} Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.710814 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.793763 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffmff\" (UniqueName: \"kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794004 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794164 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794324 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794458 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794563 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.794647 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0\") pod \"721fe1de-3687-4a8c-babb-e30f9ec46abe\" (UID: \"721fe1de-3687-4a8c-babb-e30f9ec46abe\") " Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.799193 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff" (OuterVolumeSpecName: "kube-api-access-ffmff") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "kube-api-access-ffmff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.800392 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.806144 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph" (OuterVolumeSpecName: "ceph") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.829054 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.831201 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory" (OuterVolumeSpecName: "inventory") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.832335 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.848031 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "721fe1de-3687-4a8c-babb-e30f9ec46abe" (UID: "721fe1de-3687-4a8c-babb-e30f9ec46abe"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897034 5002 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897075 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897088 5002 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897105 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffmff\" (UniqueName: \"kubernetes.io/projected/721fe1de-3687-4a8c-babb-e30f9ec46abe-kube-api-access-ffmff\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897117 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897131 5002 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:07 crc kubenswrapper[5002]: I1014 08:34:07.897144 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/721fe1de-3687-4a8c-babb-e30f9ec46abe-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.284178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" event={"ID":"721fe1de-3687-4a8c-babb-e30f9ec46abe","Type":"ContainerDied","Data":"5bab58fc20c69c3a66e174349db696b3ca284945cb9e60137a457ab9cc85c0a6"} Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.284253 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bab58fc20c69c3a66e174349db696b3ca284945cb9e60137a457ab9cc85c0a6" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.284286 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.402576 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj"] Oct 14 08:34:08 crc kubenswrapper[5002]: E1014 08:34:08.403257 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="721fe1de-3687-4a8c-babb-e30f9ec46abe" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.403276 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="721fe1de-3687-4a8c-babb-e30f9ec46abe" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.403433 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="721fe1de-3687-4a8c-babb-e30f9ec46abe" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.403987 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.407924 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.408428 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.408485 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.408434 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.411249 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.411512 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.424631 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj"] Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.508717 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.508775 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn9lt\" (UniqueName: \"kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.508894 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.508924 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.508976 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.509021 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.610740 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn9lt\" (UniqueName: \"kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.611388 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.611494 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.611598 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.611731 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.611931 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.616006 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.616221 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.617308 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.617355 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.621102 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.643202 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn9lt\" (UniqueName: \"kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:08 crc kubenswrapper[5002]: I1014 08:34:08.721033 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:34:09 crc kubenswrapper[5002]: I1014 08:34:09.123585 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj"] Oct 14 08:34:09 crc kubenswrapper[5002]: I1014 08:34:09.301916 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" event={"ID":"4d34c041-ecf1-4380-8fd9-6bd9a9660f87","Type":"ContainerStarted","Data":"cc4205e75fc29ba4e25387cb2f3a9b2542a507c8f4885ad1b2930c8fb508a7dc"} Oct 14 08:34:10 crc kubenswrapper[5002]: I1014 08:34:10.315410 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" event={"ID":"4d34c041-ecf1-4380-8fd9-6bd9a9660f87","Type":"ContainerStarted","Data":"7f88c0d9d3b3492f813c2d1e0a3eacf6e7a2b6ce561f35d68c86135980ebf474"} Oct 14 08:34:10 crc kubenswrapper[5002]: I1014 08:34:10.357915 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" podStartSLOduration=1.827643572 podStartE2EDuration="2.357866791s" podCreationTimestamp="2025-10-14 08:34:08 +0000 UTC" firstStartedPulling="2025-10-14 08:34:09.130119034 +0000 UTC m=+2582.111358506" lastFinishedPulling="2025-10-14 08:34:09.660342233 +0000 UTC m=+2582.641581725" observedRunningTime="2025-10-14 08:34:10.338937808 +0000 UTC m=+2583.320177330" watchObservedRunningTime="2025-10-14 08:34:10.357866791 +0000 UTC m=+2583.339106293" Oct 14 08:34:39 crc kubenswrapper[5002]: I1014 08:34:39.218204 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:34:39 crc kubenswrapper[5002]: I1014 08:34:39.218945 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:35:09 crc kubenswrapper[5002]: I1014 08:35:09.218038 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:35:09 crc kubenswrapper[5002]: I1014 08:35:09.218718 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:35:39 crc kubenswrapper[5002]: I1014 08:35:39.218677 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:35:39 crc kubenswrapper[5002]: I1014 08:35:39.219431 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:35:39 crc kubenswrapper[5002]: I1014 08:35:39.219503 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:35:39 crc kubenswrapper[5002]: I1014 08:35:39.220508 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:35:39 crc kubenswrapper[5002]: I1014 08:35:39.220593 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7" gracePeriod=600 Oct 14 08:35:40 crc kubenswrapper[5002]: I1014 08:35:40.363659 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7" exitCode=0 Oct 14 08:35:40 crc kubenswrapper[5002]: I1014 08:35:40.363742 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7"} Oct 14 08:35:40 crc kubenswrapper[5002]: I1014 08:35:40.364317 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca"} Oct 14 08:35:40 crc kubenswrapper[5002]: I1014 08:35:40.364339 5002 scope.go:117] "RemoveContainer" containerID="9ecb7ea6a432bf870c3118076f02742a300ac77982e74be83248520c619b6e08" Oct 14 08:35:52 crc kubenswrapper[5002]: I1014 08:35:52.974918 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:52.980299 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:52.982860 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.129814 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.130175 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phdqk\" (UniqueName: \"kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.130319 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.234180 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phdqk\" (UniqueName: \"kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.234320 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.234454 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.235270 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.235448 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.272341 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phdqk\" (UniqueName: \"kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk\") pod \"certified-operators-fdwnn\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.329266 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:35:53 crc kubenswrapper[5002]: I1014 08:35:53.827019 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:35:54 crc kubenswrapper[5002]: I1014 08:35:54.543620 5002 generic.go:334] "Generic (PLEG): container finished" podID="316eba8f-766c-49c3-b381-716be95254bb" containerID="9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a" exitCode=0 Oct 14 08:35:54 crc kubenswrapper[5002]: I1014 08:35:54.543730 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerDied","Data":"9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a"} Oct 14 08:35:54 crc kubenswrapper[5002]: I1014 08:35:54.544164 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerStarted","Data":"63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20"} Oct 14 08:35:54 crc kubenswrapper[5002]: I1014 08:35:54.546967 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:35:55 crc kubenswrapper[5002]: I1014 08:35:55.561652 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerStarted","Data":"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba"} Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.170980 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.173075 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.176399 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.293306 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.293927 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tbw8\" (UniqueName: \"kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.294231 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.396313 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tbw8\" (UniqueName: \"kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.396457 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.396642 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.397472 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.397698 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.426449 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tbw8\" (UniqueName: \"kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8\") pod \"redhat-operators-kv5pp\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.511744 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.584674 5002 generic.go:334] "Generic (PLEG): container finished" podID="316eba8f-766c-49c3-b381-716be95254bb" containerID="190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba" exitCode=0 Oct 14 08:35:56 crc kubenswrapper[5002]: I1014 08:35:56.585088 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerDied","Data":"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba"} Oct 14 08:35:57 crc kubenswrapper[5002]: W1014 08:35:57.052525 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84d3a248_e002_4f4a_9986_92ab3d5e2810.slice/crio-ac83788e27108294ca30660517c6dd6034ad25c3998a6de9d5319dec8d617235 WatchSource:0}: Error finding container ac83788e27108294ca30660517c6dd6034ad25c3998a6de9d5319dec8d617235: Status 404 returned error can't find the container with id ac83788e27108294ca30660517c6dd6034ad25c3998a6de9d5319dec8d617235 Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.052642 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.593649 5002 generic.go:334] "Generic (PLEG): container finished" podID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerID="f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a" exitCode=0 Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.593695 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerDied","Data":"f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a"} Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.594063 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerStarted","Data":"ac83788e27108294ca30660517c6dd6034ad25c3998a6de9d5319dec8d617235"} Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.596038 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerStarted","Data":"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167"} Oct 14 08:35:57 crc kubenswrapper[5002]: I1014 08:35:57.640125 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fdwnn" podStartSLOduration=3.060061647 podStartE2EDuration="5.640106734s" podCreationTimestamp="2025-10-14 08:35:52 +0000 UTC" firstStartedPulling="2025-10-14 08:35:54.546351937 +0000 UTC m=+2687.527591419" lastFinishedPulling="2025-10-14 08:35:57.126397054 +0000 UTC m=+2690.107636506" observedRunningTime="2025-10-14 08:35:57.636092127 +0000 UTC m=+2690.617331599" watchObservedRunningTime="2025-10-14 08:35:57.640106734 +0000 UTC m=+2690.621346196" Oct 14 08:35:58 crc kubenswrapper[5002]: I1014 08:35:58.609717 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerStarted","Data":"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56"} Oct 14 08:35:59 crc kubenswrapper[5002]: I1014 08:35:59.618383 5002 generic.go:334] "Generic (PLEG): container finished" podID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerID="afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56" exitCode=0 Oct 14 08:35:59 crc kubenswrapper[5002]: I1014 08:35:59.618446 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerDied","Data":"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56"} Oct 14 08:36:00 crc kubenswrapper[5002]: I1014 08:36:00.631767 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerStarted","Data":"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19"} Oct 14 08:36:00 crc kubenswrapper[5002]: I1014 08:36:00.657549 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kv5pp" podStartSLOduration=2.191560719 podStartE2EDuration="4.657525705s" podCreationTimestamp="2025-10-14 08:35:56 +0000 UTC" firstStartedPulling="2025-10-14 08:35:57.595581997 +0000 UTC m=+2690.576821459" lastFinishedPulling="2025-10-14 08:36:00.061546953 +0000 UTC m=+2693.042786445" observedRunningTime="2025-10-14 08:36:00.651090123 +0000 UTC m=+2693.632329605" watchObservedRunningTime="2025-10-14 08:36:00.657525705 +0000 UTC m=+2693.638765167" Oct 14 08:36:03 crc kubenswrapper[5002]: I1014 08:36:03.330089 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:03 crc kubenswrapper[5002]: I1014 08:36:03.330720 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:03 crc kubenswrapper[5002]: I1014 08:36:03.381482 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:03 crc kubenswrapper[5002]: I1014 08:36:03.702666 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:04 crc kubenswrapper[5002]: I1014 08:36:04.537705 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:36:05 crc kubenswrapper[5002]: I1014 08:36:05.694114 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fdwnn" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="registry-server" containerID="cri-o://166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167" gracePeriod=2 Oct 14 08:36:06 crc kubenswrapper[5002]: I1014 08:36:06.512752 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:06 crc kubenswrapper[5002]: I1014 08:36:06.513126 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:06 crc kubenswrapper[5002]: I1014 08:36:06.568649 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:06 crc kubenswrapper[5002]: I1014 08:36:06.775217 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.345211 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.412176 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities\") pod \"316eba8f-766c-49c3-b381-716be95254bb\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.412349 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content\") pod \"316eba8f-766c-49c3-b381-716be95254bb\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.412534 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phdqk\" (UniqueName: \"kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk\") pod \"316eba8f-766c-49c3-b381-716be95254bb\" (UID: \"316eba8f-766c-49c3-b381-716be95254bb\") " Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.413382 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities" (OuterVolumeSpecName: "utilities") pod "316eba8f-766c-49c3-b381-716be95254bb" (UID: "316eba8f-766c-49c3-b381-716be95254bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.420501 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk" (OuterVolumeSpecName: "kube-api-access-phdqk") pod "316eba8f-766c-49c3-b381-716be95254bb" (UID: "316eba8f-766c-49c3-b381-716be95254bb"). InnerVolumeSpecName "kube-api-access-phdqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.456038 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "316eba8f-766c-49c3-b381-716be95254bb" (UID: "316eba8f-766c-49c3-b381-716be95254bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.514475 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phdqk\" (UniqueName: \"kubernetes.io/projected/316eba8f-766c-49c3-b381-716be95254bb-kube-api-access-phdqk\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.514516 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.514530 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/316eba8f-766c-49c3-b381-716be95254bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.717529 5002 generic.go:334] "Generic (PLEG): container finished" podID="316eba8f-766c-49c3-b381-716be95254bb" containerID="166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167" exitCode=0 Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.717574 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerDied","Data":"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167"} Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.717621 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fdwnn" event={"ID":"316eba8f-766c-49c3-b381-716be95254bb","Type":"ContainerDied","Data":"63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20"} Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.717640 5002 scope.go:117] "RemoveContainer" containerID="166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.717667 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fdwnn" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.753048 5002 scope.go:117] "RemoveContainer" containerID="190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.781251 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.789031 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fdwnn"] Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.791517 5002 scope.go:117] "RemoveContainer" containerID="9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.858372 5002 scope.go:117] "RemoveContainer" containerID="166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167" Oct 14 08:36:07 crc kubenswrapper[5002]: E1014 08:36:07.870920 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167\": container with ID starting with 166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167 not found: ID does not exist" containerID="166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.871000 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167"} err="failed to get container status \"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167\": rpc error: code = NotFound desc = could not find container \"166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167\": container with ID starting with 166887fd7fced17fb744ab0eb7821062b7a8b647af78737fbd70cec64625e167 not found: ID does not exist" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.871045 5002 scope.go:117] "RemoveContainer" containerID="190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba" Oct 14 08:36:07 crc kubenswrapper[5002]: E1014 08:36:07.871638 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba\": container with ID starting with 190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba not found: ID does not exist" containerID="190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.871690 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba"} err="failed to get container status \"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba\": rpc error: code = NotFound desc = could not find container \"190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba\": container with ID starting with 190688fe38362a4c8e32fd903674e71a98988858a8d566ac781bc4ec75c810ba not found: ID does not exist" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.871725 5002 scope.go:117] "RemoveContainer" containerID="9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a" Oct 14 08:36:07 crc kubenswrapper[5002]: E1014 08:36:07.872149 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a\": container with ID starting with 9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a not found: ID does not exist" containerID="9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a" Oct 14 08:36:07 crc kubenswrapper[5002]: I1014 08:36:07.872211 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a"} err="failed to get container status \"9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a\": rpc error: code = NotFound desc = could not find container \"9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a\": container with ID starting with 9e2da0fb9b0d1602780e953c934145c4136b5e6a4b0aec65004be3260b32e88a not found: ID does not exist" Oct 14 08:36:08 crc kubenswrapper[5002]: I1014 08:36:08.341317 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:36:08 crc kubenswrapper[5002]: I1014 08:36:08.732392 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kv5pp" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="registry-server" containerID="cri-o://e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19" gracePeriod=2 Oct 14 08:36:08 crc kubenswrapper[5002]: E1014 08:36:08.802871 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache]" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.228724 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.348384 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tbw8\" (UniqueName: \"kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8\") pod \"84d3a248-e002-4f4a-9986-92ab3d5e2810\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.348749 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities\") pod \"84d3a248-e002-4f4a-9986-92ab3d5e2810\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.348805 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content\") pod \"84d3a248-e002-4f4a-9986-92ab3d5e2810\" (UID: \"84d3a248-e002-4f4a-9986-92ab3d5e2810\") " Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.349984 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities" (OuterVolumeSpecName: "utilities") pod "84d3a248-e002-4f4a-9986-92ab3d5e2810" (UID: "84d3a248-e002-4f4a-9986-92ab3d5e2810"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.359178 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8" (OuterVolumeSpecName: "kube-api-access-2tbw8") pod "84d3a248-e002-4f4a-9986-92ab3d5e2810" (UID: "84d3a248-e002-4f4a-9986-92ab3d5e2810"). InnerVolumeSpecName "kube-api-access-2tbw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.445503 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84d3a248-e002-4f4a-9986-92ab3d5e2810" (UID: "84d3a248-e002-4f4a-9986-92ab3d5e2810"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.450957 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tbw8\" (UniqueName: \"kubernetes.io/projected/84d3a248-e002-4f4a-9986-92ab3d5e2810-kube-api-access-2tbw8\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.450993 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.451009 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84d3a248-e002-4f4a-9986-92ab3d5e2810-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.745188 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="316eba8f-766c-49c3-b381-716be95254bb" path="/var/lib/kubelet/pods/316eba8f-766c-49c3-b381-716be95254bb/volumes" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.745648 5002 generic.go:334] "Generic (PLEG): container finished" podID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerID="e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19" exitCode=0 Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.745759 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kv5pp" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.749083 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerDied","Data":"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19"} Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.749171 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kv5pp" event={"ID":"84d3a248-e002-4f4a-9986-92ab3d5e2810","Type":"ContainerDied","Data":"ac83788e27108294ca30660517c6dd6034ad25c3998a6de9d5319dec8d617235"} Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.749197 5002 scope.go:117] "RemoveContainer" containerID="e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.772195 5002 scope.go:117] "RemoveContainer" containerID="afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.796137 5002 scope.go:117] "RemoveContainer" containerID="f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.802949 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.813967 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kv5pp"] Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.849331 5002 scope.go:117] "RemoveContainer" containerID="e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19" Oct 14 08:36:09 crc kubenswrapper[5002]: E1014 08:36:09.850227 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19\": container with ID starting with e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19 not found: ID does not exist" containerID="e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.850283 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19"} err="failed to get container status \"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19\": rpc error: code = NotFound desc = could not find container \"e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19\": container with ID starting with e3e1b1f4119456dcb2c62bb7114a7a570292dfc036805cc6b83b968f06446d19 not found: ID does not exist" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.850320 5002 scope.go:117] "RemoveContainer" containerID="afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56" Oct 14 08:36:09 crc kubenswrapper[5002]: E1014 08:36:09.850635 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56\": container with ID starting with afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56 not found: ID does not exist" containerID="afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.850666 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56"} err="failed to get container status \"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56\": rpc error: code = NotFound desc = could not find container \"afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56\": container with ID starting with afd93f2cf2812a8550cfc84fce5a654ac8f7a01892120041a4d59d89d9d33d56 not found: ID does not exist" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.850687 5002 scope.go:117] "RemoveContainer" containerID="f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a" Oct 14 08:36:09 crc kubenswrapper[5002]: E1014 08:36:09.850926 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a\": container with ID starting with f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a not found: ID does not exist" containerID="f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a" Oct 14 08:36:09 crc kubenswrapper[5002]: I1014 08:36:09.850970 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a"} err="failed to get container status \"f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a\": rpc error: code = NotFound desc = could not find container \"f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a\": container with ID starting with f414feff51f1a4dbbaf8703bf085b1c902f281c8f4d5b982e932dfc45ce42c0a not found: ID does not exist" Oct 14 08:36:11 crc kubenswrapper[5002]: I1014 08:36:11.742492 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" path="/var/lib/kubelet/pods/84d3a248-e002-4f4a-9986-92ab3d5e2810/volumes" Oct 14 08:36:19 crc kubenswrapper[5002]: E1014 08:36:19.048169 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache]" Oct 14 08:36:29 crc kubenswrapper[5002]: E1014 08:36:29.300816 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache]" Oct 14 08:36:39 crc kubenswrapper[5002]: E1014 08:36:39.568944 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache]" Oct 14 08:36:49 crc kubenswrapper[5002]: E1014 08:36:49.827091 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache]" Oct 14 08:37:00 crc kubenswrapper[5002]: E1014 08:37:00.075743 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice/crio-63cfa7d6aead724f94d36b03ed5d329a0fa683b19af31290c1d35ae7934dbe20\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod316eba8f_766c_49c3_b381_716be95254bb.slice\": RecentStats: unable to find data in memory cache]" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.536924 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537643 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="extract-content" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537657 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="extract-content" Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537673 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="extract-utilities" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537683 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="extract-utilities" Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537695 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537704 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537720 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537729 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537759 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="extract-content" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537766 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="extract-content" Oct 14 08:37:02 crc kubenswrapper[5002]: E1014 08:37:02.537783 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="extract-utilities" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.537791 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="extract-utilities" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.538113 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="316eba8f-766c-49c3-b381-716be95254bb" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.538136 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d3a248-e002-4f4a-9986-92ab3d5e2810" containerName="registry-server" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.539997 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.553295 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.601372 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.601452 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.601489 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf82l\" (UniqueName: \"kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.703779 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.704243 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.704278 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf82l\" (UniqueName: \"kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.704439 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.704926 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.733454 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf82l\" (UniqueName: \"kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l\") pod \"redhat-marketplace-w94hz\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:02 crc kubenswrapper[5002]: I1014 08:37:02.879228 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:03 crc kubenswrapper[5002]: I1014 08:37:03.321791 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:03 crc kubenswrapper[5002]: I1014 08:37:03.356466 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerStarted","Data":"f42e610d99b54ed23ac823c77c411920eafa1d6c7a7a9f6e6352439f1c418a42"} Oct 14 08:37:04 crc kubenswrapper[5002]: I1014 08:37:04.373355 5002 generic.go:334] "Generic (PLEG): container finished" podID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerID="b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89" exitCode=0 Oct 14 08:37:04 crc kubenswrapper[5002]: I1014 08:37:04.373424 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerDied","Data":"b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89"} Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.091619 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.094057 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.113757 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.150510 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jm5n\" (UniqueName: \"kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.150547 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.150574 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.252350 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jm5n\" (UniqueName: \"kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.252406 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.252446 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.252945 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.253051 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.282472 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jm5n\" (UniqueName: \"kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n\") pod \"community-operators-rcc2r\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.388147 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerStarted","Data":"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213"} Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.441019 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:05 crc kubenswrapper[5002]: I1014 08:37:05.992774 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:05 crc kubenswrapper[5002]: W1014 08:37:05.997763 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod065d8aaf_cbf1_416d_9afe_ee8b91704eb7.slice/crio-ffa64aa31e268869082bc14c0b4eb46f15b328763783b76da181db48f2d380ac WatchSource:0}: Error finding container ffa64aa31e268869082bc14c0b4eb46f15b328763783b76da181db48f2d380ac: Status 404 returned error can't find the container with id ffa64aa31e268869082bc14c0b4eb46f15b328763783b76da181db48f2d380ac Oct 14 08:37:06 crc kubenswrapper[5002]: I1014 08:37:06.401911 5002 generic.go:334] "Generic (PLEG): container finished" podID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerID="38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553" exitCode=0 Oct 14 08:37:06 crc kubenswrapper[5002]: I1014 08:37:06.401976 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerDied","Data":"38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553"} Oct 14 08:37:06 crc kubenswrapper[5002]: I1014 08:37:06.402341 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerStarted","Data":"ffa64aa31e268869082bc14c0b4eb46f15b328763783b76da181db48f2d380ac"} Oct 14 08:37:06 crc kubenswrapper[5002]: I1014 08:37:06.405118 5002 generic.go:334] "Generic (PLEG): container finished" podID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerID="502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213" exitCode=0 Oct 14 08:37:06 crc kubenswrapper[5002]: I1014 08:37:06.405168 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerDied","Data":"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213"} Oct 14 08:37:07 crc kubenswrapper[5002]: I1014 08:37:07.415676 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerStarted","Data":"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81"} Oct 14 08:37:07 crc kubenswrapper[5002]: I1014 08:37:07.418385 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerStarted","Data":"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6"} Oct 14 08:37:07 crc kubenswrapper[5002]: I1014 08:37:07.454300 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w94hz" podStartSLOduration=2.958794118 podStartE2EDuration="5.45428055s" podCreationTimestamp="2025-10-14 08:37:02 +0000 UTC" firstStartedPulling="2025-10-14 08:37:04.37685739 +0000 UTC m=+2757.358096842" lastFinishedPulling="2025-10-14 08:37:06.872343772 +0000 UTC m=+2759.853583274" observedRunningTime="2025-10-14 08:37:07.450029427 +0000 UTC m=+2760.431268909" watchObservedRunningTime="2025-10-14 08:37:07.45428055 +0000 UTC m=+2760.435520002" Oct 14 08:37:08 crc kubenswrapper[5002]: I1014 08:37:08.440677 5002 generic.go:334] "Generic (PLEG): container finished" podID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerID="2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81" exitCode=0 Oct 14 08:37:08 crc kubenswrapper[5002]: I1014 08:37:08.440769 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerDied","Data":"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81"} Oct 14 08:37:09 crc kubenswrapper[5002]: I1014 08:37:09.455906 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerStarted","Data":"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442"} Oct 14 08:37:09 crc kubenswrapper[5002]: I1014 08:37:09.485705 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rcc2r" podStartSLOduration=1.997466997 podStartE2EDuration="4.485685045s" podCreationTimestamp="2025-10-14 08:37:05 +0000 UTC" firstStartedPulling="2025-10-14 08:37:06.404162186 +0000 UTC m=+2759.385401648" lastFinishedPulling="2025-10-14 08:37:08.892380234 +0000 UTC m=+2761.873619696" observedRunningTime="2025-10-14 08:37:09.475187595 +0000 UTC m=+2762.456427077" watchObservedRunningTime="2025-10-14 08:37:09.485685045 +0000 UTC m=+2762.466924507" Oct 14 08:37:12 crc kubenswrapper[5002]: I1014 08:37:12.879622 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:12 crc kubenswrapper[5002]: I1014 08:37:12.882145 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:12 crc kubenswrapper[5002]: I1014 08:37:12.930047 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:13 crc kubenswrapper[5002]: I1014 08:37:13.545038 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:14 crc kubenswrapper[5002]: I1014 08:37:14.088726 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.441743 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.442920 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.496946 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.508389 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w94hz" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="registry-server" containerID="cri-o://af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6" gracePeriod=2 Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.570387 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:15 crc kubenswrapper[5002]: I1014 08:37:15.994240 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.160433 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content\") pod \"0094aa2e-0189-4456-8cb0-320563b8e18e\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.160508 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities\") pod \"0094aa2e-0189-4456-8cb0-320563b8e18e\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.160647 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf82l\" (UniqueName: \"kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l\") pod \"0094aa2e-0189-4456-8cb0-320563b8e18e\" (UID: \"0094aa2e-0189-4456-8cb0-320563b8e18e\") " Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.163068 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities" (OuterVolumeSpecName: "utilities") pod "0094aa2e-0189-4456-8cb0-320563b8e18e" (UID: "0094aa2e-0189-4456-8cb0-320563b8e18e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.171466 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l" (OuterVolumeSpecName: "kube-api-access-jf82l") pod "0094aa2e-0189-4456-8cb0-320563b8e18e" (UID: "0094aa2e-0189-4456-8cb0-320563b8e18e"). InnerVolumeSpecName "kube-api-access-jf82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.181760 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0094aa2e-0189-4456-8cb0-320563b8e18e" (UID: "0094aa2e-0189-4456-8cb0-320563b8e18e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.262725 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.262786 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0094aa2e-0189-4456-8cb0-320563b8e18e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.262813 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf82l\" (UniqueName: \"kubernetes.io/projected/0094aa2e-0189-4456-8cb0-320563b8e18e-kube-api-access-jf82l\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.527763 5002 generic.go:334] "Generic (PLEG): container finished" podID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerID="af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6" exitCode=0 Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.527885 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerDied","Data":"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6"} Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.527916 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w94hz" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.527956 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w94hz" event={"ID":"0094aa2e-0189-4456-8cb0-320563b8e18e","Type":"ContainerDied","Data":"f42e610d99b54ed23ac823c77c411920eafa1d6c7a7a9f6e6352439f1c418a42"} Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.528019 5002 scope.go:117] "RemoveContainer" containerID="af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.565801 5002 scope.go:117] "RemoveContainer" containerID="502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.580995 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.586169 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w94hz"] Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.627526 5002 scope.go:117] "RemoveContainer" containerID="b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.656618 5002 scope.go:117] "RemoveContainer" containerID="af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6" Oct 14 08:37:16 crc kubenswrapper[5002]: E1014 08:37:16.657272 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6\": container with ID starting with af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6 not found: ID does not exist" containerID="af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.657301 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6"} err="failed to get container status \"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6\": rpc error: code = NotFound desc = could not find container \"af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6\": container with ID starting with af2c7bb3dbf3e3bc96d416751048ea00a0901c0e54222d360a66c422d508d6a6 not found: ID does not exist" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.657321 5002 scope.go:117] "RemoveContainer" containerID="502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213" Oct 14 08:37:16 crc kubenswrapper[5002]: E1014 08:37:16.657788 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213\": container with ID starting with 502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213 not found: ID does not exist" containerID="502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.657810 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213"} err="failed to get container status \"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213\": rpc error: code = NotFound desc = could not find container \"502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213\": container with ID starting with 502fbb07c063dc184158c318a9626e37cb714c2808a8d9804d684448c3ab8213 not found: ID does not exist" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.657848 5002 scope.go:117] "RemoveContainer" containerID="b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89" Oct 14 08:37:16 crc kubenswrapper[5002]: E1014 08:37:16.658101 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89\": container with ID starting with b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89 not found: ID does not exist" containerID="b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.658121 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89"} err="failed to get container status \"b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89\": rpc error: code = NotFound desc = could not find container \"b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89\": container with ID starting with b4b25883e077aafd472444230036357c6ffb027e2cfda8962fa658c229ee9f89 not found: ID does not exist" Oct 14 08:37:16 crc kubenswrapper[5002]: I1014 08:37:16.889906 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:17 crc kubenswrapper[5002]: I1014 08:37:17.536006 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rcc2r" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="registry-server" containerID="cri-o://241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442" gracePeriod=2 Oct 14 08:37:17 crc kubenswrapper[5002]: I1014 08:37:17.778471 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" path="/var/lib/kubelet/pods/0094aa2e-0189-4456-8cb0-320563b8e18e/volumes" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.087719 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.196559 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities\") pod \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.196709 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content\") pod \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.196931 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jm5n\" (UniqueName: \"kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n\") pod \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\" (UID: \"065d8aaf-cbf1-416d-9afe-ee8b91704eb7\") " Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.197312 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities" (OuterVolumeSpecName: "utilities") pod "065d8aaf-cbf1-416d-9afe-ee8b91704eb7" (UID: "065d8aaf-cbf1-416d-9afe-ee8b91704eb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.203019 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n" (OuterVolumeSpecName: "kube-api-access-2jm5n") pod "065d8aaf-cbf1-416d-9afe-ee8b91704eb7" (UID: "065d8aaf-cbf1-416d-9afe-ee8b91704eb7"). InnerVolumeSpecName "kube-api-access-2jm5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.272024 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "065d8aaf-cbf1-416d-9afe-ee8b91704eb7" (UID: "065d8aaf-cbf1-416d-9afe-ee8b91704eb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.303082 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.303125 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jm5n\" (UniqueName: \"kubernetes.io/projected/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-kube-api-access-2jm5n\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.303138 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065d8aaf-cbf1-416d-9afe-ee8b91704eb7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.544531 5002 generic.go:334] "Generic (PLEG): container finished" podID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerID="241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442" exitCode=0 Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.544589 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerDied","Data":"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442"} Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.544601 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rcc2r" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.544635 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rcc2r" event={"ID":"065d8aaf-cbf1-416d-9afe-ee8b91704eb7","Type":"ContainerDied","Data":"ffa64aa31e268869082bc14c0b4eb46f15b328763783b76da181db48f2d380ac"} Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.544664 5002 scope.go:117] "RemoveContainer" containerID="241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.564505 5002 scope.go:117] "RemoveContainer" containerID="2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.582042 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.587986 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rcc2r"] Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.603586 5002 scope.go:117] "RemoveContainer" containerID="38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.632136 5002 scope.go:117] "RemoveContainer" containerID="241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442" Oct 14 08:37:18 crc kubenswrapper[5002]: E1014 08:37:18.632555 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442\": container with ID starting with 241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442 not found: ID does not exist" containerID="241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.632596 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442"} err="failed to get container status \"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442\": rpc error: code = NotFound desc = could not find container \"241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442\": container with ID starting with 241e0e0d5858861bf886acea47160d2dfef4b3649ddf464274a510c8c6e67442 not found: ID does not exist" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.632621 5002 scope.go:117] "RemoveContainer" containerID="2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81" Oct 14 08:37:18 crc kubenswrapper[5002]: E1014 08:37:18.633146 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81\": container with ID starting with 2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81 not found: ID does not exist" containerID="2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.633175 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81"} err="failed to get container status \"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81\": rpc error: code = NotFound desc = could not find container \"2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81\": container with ID starting with 2ef006180076cfdb6feb0210442eaaf5ee6998c610c85067aa27d7b9a40d6b81 not found: ID does not exist" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.633194 5002 scope.go:117] "RemoveContainer" containerID="38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553" Oct 14 08:37:18 crc kubenswrapper[5002]: E1014 08:37:18.633443 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553\": container with ID starting with 38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553 not found: ID does not exist" containerID="38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553" Oct 14 08:37:18 crc kubenswrapper[5002]: I1014 08:37:18.633477 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553"} err="failed to get container status \"38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553\": rpc error: code = NotFound desc = could not find container \"38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553\": container with ID starting with 38f433dbe617a7f4506c510fb2bbf8e4bb45d8a33add304c28d3abb22273e553 not found: ID does not exist" Oct 14 08:37:19 crc kubenswrapper[5002]: I1014 08:37:19.740589 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" path="/var/lib/kubelet/pods/065d8aaf-cbf1-416d-9afe-ee8b91704eb7/volumes" Oct 14 08:37:39 crc kubenswrapper[5002]: I1014 08:37:39.218176 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:37:39 crc kubenswrapper[5002]: I1014 08:37:39.220161 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:38:09 crc kubenswrapper[5002]: I1014 08:38:09.218234 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:38:09 crc kubenswrapper[5002]: I1014 08:38:09.220572 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.218578 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.219258 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.219338 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.220286 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.220381 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" gracePeriod=600 Oct 14 08:38:39 crc kubenswrapper[5002]: E1014 08:38:39.353690 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.473747 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" exitCode=0 Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.473812 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca"} Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.473908 5002 scope.go:117] "RemoveContainer" containerID="602d8664c4e21db79100a9a88f5998294092472f94dfdf62df4c6c4f46ad68c7" Oct 14 08:38:39 crc kubenswrapper[5002]: I1014 08:38:39.474633 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:38:39 crc kubenswrapper[5002]: E1014 08:38:39.477247 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:38:51 crc kubenswrapper[5002]: I1014 08:38:51.721365 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:38:51 crc kubenswrapper[5002]: E1014 08:38:51.722518 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:38:57 crc kubenswrapper[5002]: I1014 08:38:57.680335 5002 generic.go:334] "Generic (PLEG): container finished" podID="4d34c041-ecf1-4380-8fd9-6bd9a9660f87" containerID="7f88c0d9d3b3492f813c2d1e0a3eacf6e7a2b6ce561f35d68c86135980ebf474" exitCode=0 Oct 14 08:38:57 crc kubenswrapper[5002]: I1014 08:38:57.680446 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" event={"ID":"4d34c041-ecf1-4380-8fd9-6bd9a9660f87","Type":"ContainerDied","Data":"7f88c0d9d3b3492f813c2d1e0a3eacf6e7a2b6ce561f35d68c86135980ebf474"} Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.189249 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242177 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242244 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242320 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242375 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn9lt\" (UniqueName: \"kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242595 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.242633 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0\") pod \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\" (UID: \"4d34c041-ecf1-4380-8fd9-6bd9a9660f87\") " Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.248233 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph" (OuterVolumeSpecName: "ceph") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.250290 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt" (OuterVolumeSpecName: "kube-api-access-fn9lt") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "kube-api-access-fn9lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.250624 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.285470 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory" (OuterVolumeSpecName: "inventory") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.289419 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.289901 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4d34c041-ecf1-4380-8fd9-6bd9a9660f87" (UID: "4d34c041-ecf1-4380-8fd9-6bd9a9660f87"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345487 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345538 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn9lt\" (UniqueName: \"kubernetes.io/projected/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-kube-api-access-fn9lt\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345552 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345561 5002 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345570 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.345581 5002 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d34c041-ecf1-4380-8fd9-6bd9a9660f87-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.716820 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" event={"ID":"4d34c041-ecf1-4380-8fd9-6bd9a9660f87","Type":"ContainerDied","Data":"cc4205e75fc29ba4e25387cb2f3a9b2542a507c8f4885ad1b2930c8fb508a7dc"} Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.717346 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc4205e75fc29ba4e25387cb2f3a9b2542a507c8f4885ad1b2930c8fb508a7dc" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.717012 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.844976 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5"] Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845327 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845341 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845357 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="extract-utilities" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845363 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="extract-utilities" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845377 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d34c041-ecf1-4380-8fd9-6bd9a9660f87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845384 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d34c041-ecf1-4380-8fd9-6bd9a9660f87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845404 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="extract-utilities" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845410 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="extract-utilities" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845430 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="extract-content" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845436 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="extract-content" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845444 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845449 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: E1014 08:38:59.845458 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="extract-content" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845464 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="extract-content" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845627 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="065d8aaf-cbf1-416d-9afe-ee8b91704eb7" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845638 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d34c041-ecf1-4380-8fd9-6bd9a9660f87" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.845651 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0094aa2e-0189-4456-8cb0-320563b8e18e" containerName="registry-server" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.846265 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.849386 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.851397 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.852509 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.852562 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.855016 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.855094 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m8f28" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.855014 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.856624 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.861546 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.874527 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5"] Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.958683 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.958790 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.958859 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldp6\" (UniqueName: \"kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.959203 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.959286 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.959562 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.959657 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.959798 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.961023 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.961210 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:38:59 crc kubenswrapper[5002]: I1014 08:38:59.961243 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.062925 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.063285 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.063687 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.063810 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.063935 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldp6\" (UniqueName: \"kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064031 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064122 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064311 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064407 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064540 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.064653 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.065385 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.066010 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.068658 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.068827 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.069337 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.069947 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.071703 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.073215 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.073723 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.074290 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.084082 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldp6\" (UniqueName: \"kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.167473 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.526287 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5"] Oct 14 08:39:00 crc kubenswrapper[5002]: W1014 08:39:00.530522 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99e9a85b_e7b7_4f1d_89e8_ea6b5c86efb2.slice/crio-d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1 WatchSource:0}: Error finding container d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1: Status 404 returned error can't find the container with id d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1 Oct 14 08:39:00 crc kubenswrapper[5002]: I1014 08:39:00.731681 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" event={"ID":"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2","Type":"ContainerStarted","Data":"d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1"} Oct 14 08:39:01 crc kubenswrapper[5002]: I1014 08:39:01.745257 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" event={"ID":"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2","Type":"ContainerStarted","Data":"e87e9a9d9870fc4ea2bac9602da34a8b5de29dc0ab73b3ea0c4187b4300f61db"} Oct 14 08:39:01 crc kubenswrapper[5002]: I1014 08:39:01.772039 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" podStartSLOduration=2.28543925 podStartE2EDuration="2.772010647s" podCreationTimestamp="2025-10-14 08:38:59 +0000 UTC" firstStartedPulling="2025-10-14 08:39:00.534595401 +0000 UTC m=+2873.515834843" lastFinishedPulling="2025-10-14 08:39:01.021166738 +0000 UTC m=+2874.002406240" observedRunningTime="2025-10-14 08:39:01.771600796 +0000 UTC m=+2874.752840318" watchObservedRunningTime="2025-10-14 08:39:01.772010647 +0000 UTC m=+2874.753250129" Oct 14 08:39:03 crc kubenswrapper[5002]: I1014 08:39:03.721355 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:39:03 crc kubenswrapper[5002]: E1014 08:39:03.722250 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:39:15 crc kubenswrapper[5002]: I1014 08:39:15.720555 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:39:15 crc kubenswrapper[5002]: E1014 08:39:15.721421 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:39:29 crc kubenswrapper[5002]: I1014 08:39:29.721076 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:39:29 crc kubenswrapper[5002]: E1014 08:39:29.722317 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:39:42 crc kubenswrapper[5002]: I1014 08:39:42.721095 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:39:42 crc kubenswrapper[5002]: E1014 08:39:42.722823 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:39:54 crc kubenswrapper[5002]: I1014 08:39:54.720875 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:39:54 crc kubenswrapper[5002]: E1014 08:39:54.722090 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:40:05 crc kubenswrapper[5002]: I1014 08:40:05.720186 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:40:05 crc kubenswrapper[5002]: E1014 08:40:05.720882 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:40:16 crc kubenswrapper[5002]: I1014 08:40:16.721254 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:40:16 crc kubenswrapper[5002]: E1014 08:40:16.722461 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:40:27 crc kubenswrapper[5002]: I1014 08:40:27.720757 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:40:27 crc kubenswrapper[5002]: E1014 08:40:27.721775 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:40:41 crc kubenswrapper[5002]: I1014 08:40:41.720565 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:40:41 crc kubenswrapper[5002]: E1014 08:40:41.721519 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:40:54 crc kubenswrapper[5002]: I1014 08:40:54.721451 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:40:54 crc kubenswrapper[5002]: E1014 08:40:54.722580 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:41:05 crc kubenswrapper[5002]: I1014 08:41:05.721183 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:41:05 crc kubenswrapper[5002]: E1014 08:41:05.722256 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:41:20 crc kubenswrapper[5002]: I1014 08:41:20.721506 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:41:20 crc kubenswrapper[5002]: E1014 08:41:20.722692 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:41:32 crc kubenswrapper[5002]: I1014 08:41:32.720709 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:41:32 crc kubenswrapper[5002]: E1014 08:41:32.721695 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:41:44 crc kubenswrapper[5002]: I1014 08:41:44.720584 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:41:44 crc kubenswrapper[5002]: E1014 08:41:44.721573 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:41:56 crc kubenswrapper[5002]: I1014 08:41:56.721413 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:41:56 crc kubenswrapper[5002]: E1014 08:41:56.722567 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:42:11 crc kubenswrapper[5002]: I1014 08:42:11.720283 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:42:11 crc kubenswrapper[5002]: E1014 08:42:11.721561 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:42:22 crc kubenswrapper[5002]: I1014 08:42:22.720524 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:42:22 crc kubenswrapper[5002]: E1014 08:42:22.721427 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:42:35 crc kubenswrapper[5002]: I1014 08:42:35.720402 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:42:35 crc kubenswrapper[5002]: E1014 08:42:35.721237 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:42:50 crc kubenswrapper[5002]: I1014 08:42:50.720425 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:42:50 crc kubenswrapper[5002]: E1014 08:42:50.721278 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:43:01 crc kubenswrapper[5002]: I1014 08:43:01.721116 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:43:01 crc kubenswrapper[5002]: E1014 08:43:01.722418 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:43:04 crc kubenswrapper[5002]: I1014 08:43:04.564606 5002 generic.go:334] "Generic (PLEG): container finished" podID="99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" containerID="e87e9a9d9870fc4ea2bac9602da34a8b5de29dc0ab73b3ea0c4187b4300f61db" exitCode=0 Oct 14 08:43:04 crc kubenswrapper[5002]: I1014 08:43:04.564693 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" event={"ID":"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2","Type":"ContainerDied","Data":"e87e9a9d9870fc4ea2bac9602da34a8b5de29dc0ab73b3ea0c4187b4300f61db"} Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.086516 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.192197 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.192249 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.192303 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.192334 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.192384 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193161 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193196 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193221 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bldp6\" (UniqueName: \"kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193256 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193316 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.193345 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.198531 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6" (OuterVolumeSpecName: "kube-api-access-bldp6") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "kube-api-access-bldp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.202120 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph" (OuterVolumeSpecName: "ceph") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.202225 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.217275 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.225223 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: E1014 08:43:06.245110 5002 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1 podName:99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2 nodeName:}" failed. No retries permitted until 2025-10-14 08:43:06.745077953 +0000 UTC m=+3119.726317415 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "nova-cell1-compute-config-1" (UniqueName: "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2") : error deleting /var/lib/kubelet/pods/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2/volume-subpaths: remove /var/lib/kubelet/pods/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2/volume-subpaths: no such file or directory Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.246155 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.246704 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.247369 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.248578 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory" (OuterVolumeSpecName: "inventory") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.250905 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295247 5002 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295280 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bldp6\" (UniqueName: \"kubernetes.io/projected/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-kube-api-access-bldp6\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295290 5002 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295299 5002 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295308 5002 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295316 5002 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295326 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295335 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295344 5002 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.295352 5002 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.588375 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" event={"ID":"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2","Type":"ContainerDied","Data":"d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1"} Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.588438 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d867f31ca861d556a0d366b664df2f2d53503e44081c1b755292432eda07b8f1" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.588523 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.803604 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") pod \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\" (UID: \"99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2\") " Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.808774 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" (UID: "99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:06 crc kubenswrapper[5002]: I1014 08:43:06.916368 5002 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:13 crc kubenswrapper[5002]: I1014 08:43:13.721394 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:43:13 crc kubenswrapper[5002]: E1014 08:43:13.722345 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.763750 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 08:43:20 crc kubenswrapper[5002]: E1014 08:43:20.765596 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.765784 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.766046 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.767068 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.769815 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.770164 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.775451 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.845915 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.847496 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.849149 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.862975 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905205 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905418 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905459 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905535 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905593 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905616 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905661 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905687 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905724 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905780 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905798 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz4vk\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-kube-api-access-tz4vk\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905831 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905872 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905944 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.905993 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:20 crc kubenswrapper[5002]: I1014 08:43:20.906020 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008075 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008124 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008146 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008171 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-lib-modules\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008190 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008330 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f22w\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-kube-api-access-4f22w\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008354 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008375 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008390 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-run\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008406 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz4vk\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-kube-api-access-tz4vk\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008431 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008448 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008471 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008494 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008507 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data-custom\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008526 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008551 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008567 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-ceph\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008570 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008603 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008700 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008760 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-dev\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008850 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008897 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-scripts\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008935 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008955 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.008976 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009020 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-nvme\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009085 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009117 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009147 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-sys\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009194 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009220 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009448 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009490 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009635 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009791 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009913 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009949 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009975 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.009994 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.010331 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.018319 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.018725 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.029573 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.030126 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.036428 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.036716 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz4vk\" (UniqueName: \"kubernetes.io/projected/a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a-kube-api-access-tz4vk\") pod \"cinder-volume-volume1-0\" (UID: \"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a\") " pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.092321 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111036 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111095 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-scripts\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111120 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111152 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-nvme\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111196 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111219 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-sys\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111258 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111258 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111287 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-lib-modules\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111321 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f22w\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-kube-api-access-4f22w\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111329 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111345 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111368 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-run\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111373 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-nvme\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111406 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-run\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111418 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111441 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-sys\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111455 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111469 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111475 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data-custom\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111499 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-lib-modules\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111522 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-ceph\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111574 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-dev\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111658 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-dev\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111702 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.111731 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/947933b0-5baf-4203-bea2-3b229c31ab91-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.114569 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-scripts\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.115079 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.115829 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.129381 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-ceph\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.130495 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/947933b0-5baf-4203-bea2-3b229c31ab91-config-data-custom\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.146666 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f22w\" (UniqueName: \"kubernetes.io/projected/947933b0-5baf-4203-bea2-3b229c31ab91-kube-api-access-4f22w\") pod \"cinder-backup-0\" (UID: \"947933b0-5baf-4203-bea2-3b229c31ab91\") " pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.166231 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.229635 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-qqxjr"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.231990 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.237853 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-qqxjr"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.319793 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67d69\" (UniqueName: \"kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69\") pod \"manila-db-create-qqxjr\" (UID: \"128b8127-99a6-4ffb-98fa-cff179c1ec85\") " pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.421717 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67d69\" (UniqueName: \"kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69\") pod \"manila-db-create-qqxjr\" (UID: \"128b8127-99a6-4ffb-98fa-cff179c1ec85\") " pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.440803 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67d69\" (UniqueName: \"kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69\") pod \"manila-db-create-qqxjr\" (UID: \"128b8127-99a6-4ffb-98fa-cff179c1ec85\") " pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.577786 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.630762 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.632770 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.635761 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.635849 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.636093 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.640998 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xvzng" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.642182 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.692059 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.706120 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.706887 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.707643 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.712414 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.712632 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.715501 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.734808 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.734881 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.734902 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.734937 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bzc\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-kube-api-access-m7bzc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.734976 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-ceph\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.735011 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-logs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.735030 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.735063 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.735107 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.775178 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a","Type":"ContainerStarted","Data":"a0469f91870b56d3b602483386818a0cc7fb881c3e41bee8483d1a73c3a48021"} Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.782721 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 14 08:43:21 crc kubenswrapper[5002]: W1014 08:43:21.784977 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod947933b0_5baf_4203_bea2_3b229c31ab91.slice/crio-21aba2a3f8df59f6cc2b351beec78d3036dbbd187c7f4cdc82ce2fdc6d6364b1 WatchSource:0}: Error finding container 21aba2a3f8df59f6cc2b351beec78d3036dbbd187c7f4cdc82ce2fdc6d6364b1: Status 404 returned error can't find the container with id 21aba2a3f8df59f6cc2b351beec78d3036dbbd187c7f4cdc82ce2fdc6d6364b1 Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836142 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836180 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836212 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836240 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7bzc\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-kube-api-access-m7bzc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836261 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836518 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-ceph\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836665 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv59s\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-kube-api-access-vv59s\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836694 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836709 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836738 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836776 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-logs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836805 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836844 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836876 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836921 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836941 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836958 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.836975 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.837009 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.837314 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.838437 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b995566-e3f1-4c1c-a9e0-02874deed49b-logs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.841362 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.841567 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.842156 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-ceph\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.842633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.844048 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b995566-e3f1-4c1c-a9e0-02874deed49b-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.855598 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7bzc\" (UniqueName: \"kubernetes.io/projected/0b995566-e3f1-4c1c-a9e0-02874deed49b-kube-api-access-m7bzc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.872503 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"0b995566-e3f1-4c1c-a9e0-02874deed49b\") " pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.939098 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv59s\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-kube-api-access-vv59s\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.940121 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.940628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.941424 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.941638 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.941780 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.941961 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.942093 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.942199 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.943057 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.940587 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f64389b8-a21d-4cf5-abe8-402b170db4b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.948978 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.949218 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.949243 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.949645 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.949868 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.950459 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f64389b8-a21d-4cf5-abe8-402b170db4b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.955524 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv59s\" (UniqueName: \"kubernetes.io/projected/f64389b8-a21d-4cf5-abe8-402b170db4b3-kube-api-access-vv59s\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.962610 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:43:21 crc kubenswrapper[5002]: I1014 08:43:21.984744 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"f64389b8-a21d-4cf5-abe8-402b170db4b3\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.027070 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.040561 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-qqxjr"] Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.437237 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:43:22 crc kubenswrapper[5002]: W1014 08:43:22.449391 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf64389b8_a21d_4cf5_abe8_402b170db4b3.slice/crio-aaaffc42a41c1431fbe301ec767006f47f0201867b2a58e876bcd051af5178a9 WatchSource:0}: Error finding container aaaffc42a41c1431fbe301ec767006f47f0201867b2a58e876bcd051af5178a9: Status 404 returned error can't find the container with id aaaffc42a41c1431fbe301ec767006f47f0201867b2a58e876bcd051af5178a9 Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.572419 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:43:22 crc kubenswrapper[5002]: W1014 08:43:22.645086 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b995566_e3f1_4c1c_a9e0_02874deed49b.slice/crio-bb58d01ed659bdadc1e76dbefd225dc9ebd326c467ae73777c3631b3fea721d3 WatchSource:0}: Error finding container bb58d01ed659bdadc1e76dbefd225dc9ebd326c467ae73777c3631b3fea721d3: Status 404 returned error can't find the container with id bb58d01ed659bdadc1e76dbefd225dc9ebd326c467ae73777c3631b3fea721d3 Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.788725 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f64389b8-a21d-4cf5-abe8-402b170db4b3","Type":"ContainerStarted","Data":"aaaffc42a41c1431fbe301ec767006f47f0201867b2a58e876bcd051af5178a9"} Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.789923 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b995566-e3f1-4c1c-a9e0-02874deed49b","Type":"ContainerStarted","Data":"bb58d01ed659bdadc1e76dbefd225dc9ebd326c467ae73777c3631b3fea721d3"} Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.816055 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"947933b0-5baf-4203-bea2-3b229c31ab91","Type":"ContainerStarted","Data":"21aba2a3f8df59f6cc2b351beec78d3036dbbd187c7f4cdc82ce2fdc6d6364b1"} Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.840410 5002 generic.go:334] "Generic (PLEG): container finished" podID="128b8127-99a6-4ffb-98fa-cff179c1ec85" containerID="b793c0a550919208c0c7559888c2a04e211cbbdfa51bba65dfbbed6d5ee2b39f" exitCode=0 Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.840681 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-qqxjr" event={"ID":"128b8127-99a6-4ffb-98fa-cff179c1ec85","Type":"ContainerDied","Data":"b793c0a550919208c0c7559888c2a04e211cbbdfa51bba65dfbbed6d5ee2b39f"} Oct 14 08:43:22 crc kubenswrapper[5002]: I1014 08:43:22.840707 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-qqxjr" event={"ID":"128b8127-99a6-4ffb-98fa-cff179c1ec85","Type":"ContainerStarted","Data":"5ca0c09d10d153d77b894de17d2718f46a38489bc0f0615be5f9efabaece4b38"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.849119 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"947933b0-5baf-4203-bea2-3b229c31ab91","Type":"ContainerStarted","Data":"58b30e27317d21f4235d9461d7a65aac5d5dcb674742693ab1ca0b822a5bc16c"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.849364 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"947933b0-5baf-4203-bea2-3b229c31ab91","Type":"ContainerStarted","Data":"3249b64aef9ae1e46e9eb01ac1101c7f997e33bbd4f63da91e4d1f43f7dc7805"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.863232 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a","Type":"ContainerStarted","Data":"cc40a16cf75855e6d371c9a06880b3effa1dac1545988da15b3b2ddcd0b17fdf"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.863269 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a","Type":"ContainerStarted","Data":"1dd1173fe8007d1ebc560b49bb7e6978c35c20080bd27dec3da7f697d7224de6"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.866329 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f64389b8-a21d-4cf5-abe8-402b170db4b3","Type":"ContainerStarted","Data":"09514e7f46185fe47996ac30b3761a50a9758efe2224b4e3bd234bbe087a0f4b"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.869932 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b995566-e3f1-4c1c-a9e0-02874deed49b","Type":"ContainerStarted","Data":"9e81511e223ed2e2b3d892d31a0ce843996b49bb99f6409ea05d96562ac27ffc"} Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.884365 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.724520171 podStartE2EDuration="3.884349451s" podCreationTimestamp="2025-10-14 08:43:20 +0000 UTC" firstStartedPulling="2025-10-14 08:43:21.786845694 +0000 UTC m=+3134.768085146" lastFinishedPulling="2025-10-14 08:43:22.946674974 +0000 UTC m=+3135.927914426" observedRunningTime="2025-10-14 08:43:23.877627291 +0000 UTC m=+3136.858866763" watchObservedRunningTime="2025-10-14 08:43:23.884349451 +0000 UTC m=+3136.865588893" Oct 14 08:43:23 crc kubenswrapper[5002]: I1014 08:43:23.914424 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.901331897 podStartE2EDuration="3.914403766s" podCreationTimestamp="2025-10-14 08:43:20 +0000 UTC" firstStartedPulling="2025-10-14 08:43:21.706669754 +0000 UTC m=+3134.687909206" lastFinishedPulling="2025-10-14 08:43:22.719741623 +0000 UTC m=+3135.700981075" observedRunningTime="2025-10-14 08:43:23.905670352 +0000 UTC m=+3136.886909804" watchObservedRunningTime="2025-10-14 08:43:23.914403766 +0000 UTC m=+3136.895643218" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.266685 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.404972 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67d69\" (UniqueName: \"kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69\") pod \"128b8127-99a6-4ffb-98fa-cff179c1ec85\" (UID: \"128b8127-99a6-4ffb-98fa-cff179c1ec85\") " Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.420048 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69" (OuterVolumeSpecName: "kube-api-access-67d69") pod "128b8127-99a6-4ffb-98fa-cff179c1ec85" (UID: "128b8127-99a6-4ffb-98fa-cff179c1ec85"). InnerVolumeSpecName "kube-api-access-67d69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.507382 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67d69\" (UniqueName: \"kubernetes.io/projected/128b8127-99a6-4ffb-98fa-cff179c1ec85-kube-api-access-67d69\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.882206 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-qqxjr" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.882193 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-qqxjr" event={"ID":"128b8127-99a6-4ffb-98fa-cff179c1ec85","Type":"ContainerDied","Data":"5ca0c09d10d153d77b894de17d2718f46a38489bc0f0615be5f9efabaece4b38"} Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.882357 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca0c09d10d153d77b894de17d2718f46a38489bc0f0615be5f9efabaece4b38" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.888808 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f64389b8-a21d-4cf5-abe8-402b170db4b3","Type":"ContainerStarted","Data":"0669869042c514a8e001366193d89bfbf1b917aa5c2c07fd61ad33e0832035ea"} Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.893072 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b995566-e3f1-4c1c-a9e0-02874deed49b","Type":"ContainerStarted","Data":"a0e0d4d0b91ac2ff71a72fed36bd31121d55ef08a8d2c5c8dafd75b36e153396"} Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.922576 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.922558852 podStartE2EDuration="4.922558852s" podCreationTimestamp="2025-10-14 08:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:43:24.918469143 +0000 UTC m=+3137.899708635" watchObservedRunningTime="2025-10-14 08:43:24.922558852 +0000 UTC m=+3137.903798304" Oct 14 08:43:24 crc kubenswrapper[5002]: I1014 08:43:24.957024 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.956987865 podStartE2EDuration="4.956987865s" podCreationTimestamp="2025-10-14 08:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:43:24.956115282 +0000 UTC m=+3137.937354734" watchObservedRunningTime="2025-10-14 08:43:24.956987865 +0000 UTC m=+3137.938227317" Oct 14 08:43:26 crc kubenswrapper[5002]: I1014 08:43:26.093078 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:26 crc kubenswrapper[5002]: I1014 08:43:26.167419 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 14 08:43:27 crc kubenswrapper[5002]: I1014 08:43:27.733798 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:43:27 crc kubenswrapper[5002]: E1014 08:43:27.734986 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.317179 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-9b88-account-create-n5r4b"] Oct 14 08:43:31 crc kubenswrapper[5002]: E1014 08:43:31.318128 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="128b8127-99a6-4ffb-98fa-cff179c1ec85" containerName="mariadb-database-create" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.318141 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="128b8127-99a6-4ffb-98fa-cff179c1ec85" containerName="mariadb-database-create" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.318354 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="128b8127-99a6-4ffb-98fa-cff179c1ec85" containerName="mariadb-database-create" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.319070 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.321452 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.342891 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9b88-account-create-n5r4b"] Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.346509 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.394072 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x57kz\" (UniqueName: \"kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz\") pod \"manila-9b88-account-create-n5r4b\" (UID: \"569d4442-3987-42ba-b796-286146ce2cdc\") " pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.450541 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.496234 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x57kz\" (UniqueName: \"kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz\") pod \"manila-9b88-account-create-n5r4b\" (UID: \"569d4442-3987-42ba-b796-286146ce2cdc\") " pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.519860 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x57kz\" (UniqueName: \"kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz\") pod \"manila-9b88-account-create-n5r4b\" (UID: \"569d4442-3987-42ba-b796-286146ce2cdc\") " pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.644182 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.963209 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.963477 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:43:31 crc kubenswrapper[5002]: I1014 08:43:31.990211 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-9b88-account-create-n5r4b"] Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.016437 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.027857 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.027910 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.031928 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.066936 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.134912 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.973361 5002 generic.go:334] "Generic (PLEG): container finished" podID="569d4442-3987-42ba-b796-286146ce2cdc" containerID="3a1f10ca2b6516e7a18ac25d93be6b9cb1770d1fe2c0c7cb124f8f64bb6530e2" exitCode=0 Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.973583 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9b88-account-create-n5r4b" event={"ID":"569d4442-3987-42ba-b796-286146ce2cdc","Type":"ContainerDied","Data":"3a1f10ca2b6516e7a18ac25d93be6b9cb1770d1fe2c0c7cb124f8f64bb6530e2"} Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.974179 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9b88-account-create-n5r4b" event={"ID":"569d4442-3987-42ba-b796-286146ce2cdc","Type":"ContainerStarted","Data":"cb250385c17467c6cb60a57db0535181f411b921f5f2124e3cade9d86011066e"} Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.974956 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.975005 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.975027 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:43:32 crc kubenswrapper[5002]: I1014 08:43:32.975053 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:43:34 crc kubenswrapper[5002]: I1014 08:43:34.431185 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:34 crc kubenswrapper[5002]: I1014 08:43:34.466552 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x57kz\" (UniqueName: \"kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz\") pod \"569d4442-3987-42ba-b796-286146ce2cdc\" (UID: \"569d4442-3987-42ba-b796-286146ce2cdc\") " Oct 14 08:43:34 crc kubenswrapper[5002]: I1014 08:43:34.472162 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz" (OuterVolumeSpecName: "kube-api-access-x57kz") pod "569d4442-3987-42ba-b796-286146ce2cdc" (UID: "569d4442-3987-42ba-b796-286146ce2cdc"). InnerVolumeSpecName "kube-api-access-x57kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:43:34 crc kubenswrapper[5002]: I1014 08:43:34.568924 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x57kz\" (UniqueName: \"kubernetes.io/projected/569d4442-3987-42ba-b796-286146ce2cdc-kube-api-access-x57kz\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.010803 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-9b88-account-create-n5r4b" event={"ID":"569d4442-3987-42ba-b796-286146ce2cdc","Type":"ContainerDied","Data":"cb250385c17467c6cb60a57db0535181f411b921f5f2124e3cade9d86011066e"} Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.011217 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb250385c17467c6cb60a57db0535181f411b921f5f2124e3cade9d86011066e" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.010855 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-9b88-account-create-n5r4b" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.010855 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.011584 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.011425 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.011808 5002 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.121597 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.267812 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:35 crc kubenswrapper[5002]: I1014 08:43:35.278020 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.590108 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-ddkcm"] Oct 14 08:43:36 crc kubenswrapper[5002]: E1014 08:43:36.590774 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="569d4442-3987-42ba-b796-286146ce2cdc" containerName="mariadb-account-create" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.590786 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="569d4442-3987-42ba-b796-286146ce2cdc" containerName="mariadb-account-create" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.590972 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="569d4442-3987-42ba-b796-286146ce2cdc" containerName="mariadb-account-create" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.592655 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.595374 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-lcb82" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.605377 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ddkcm"] Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.606177 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.709087 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bxm9\" (UniqueName: \"kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.709487 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.709531 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.709583 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.810865 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.810930 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.810974 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.811304 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bxm9\" (UniqueName: \"kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.818774 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.820171 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.823215 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.833564 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bxm9\" (UniqueName: \"kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9\") pod \"manila-db-sync-ddkcm\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:36 crc kubenswrapper[5002]: I1014 08:43:36.923389 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:37 crc kubenswrapper[5002]: I1014 08:43:37.468438 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ddkcm"] Oct 14 08:43:38 crc kubenswrapper[5002]: I1014 08:43:38.057719 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ddkcm" event={"ID":"431c8c01-259f-4ed7-8244-f16a549042ef","Type":"ContainerStarted","Data":"eca1f6ef4e7f6e9b5c34548a5b7825c4a62bc28603b16446a712bbff55c668b4"} Oct 14 08:43:40 crc kubenswrapper[5002]: I1014 08:43:40.720563 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:43:43 crc kubenswrapper[5002]: I1014 08:43:43.124259 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a"} Oct 14 08:43:43 crc kubenswrapper[5002]: I1014 08:43:43.126487 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ddkcm" event={"ID":"431c8c01-259f-4ed7-8244-f16a549042ef","Type":"ContainerStarted","Data":"f4774067f075249866c85c6b565ab3c2fd9a387c84e9810accf0a1e09da08a54"} Oct 14 08:43:43 crc kubenswrapper[5002]: I1014 08:43:43.170188 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-ddkcm" podStartSLOduration=2.027445989 podStartE2EDuration="7.170172971s" podCreationTimestamp="2025-10-14 08:43:36 +0000 UTC" firstStartedPulling="2025-10-14 08:43:37.476740501 +0000 UTC m=+3150.457979953" lastFinishedPulling="2025-10-14 08:43:42.619467483 +0000 UTC m=+3155.600706935" observedRunningTime="2025-10-14 08:43:43.165153766 +0000 UTC m=+3156.146393218" watchObservedRunningTime="2025-10-14 08:43:43.170172971 +0000 UTC m=+3156.151412423" Oct 14 08:43:53 crc kubenswrapper[5002]: I1014 08:43:53.232730 5002 generic.go:334] "Generic (PLEG): container finished" podID="431c8c01-259f-4ed7-8244-f16a549042ef" containerID="f4774067f075249866c85c6b565ab3c2fd9a387c84e9810accf0a1e09da08a54" exitCode=0 Oct 14 08:43:53 crc kubenswrapper[5002]: I1014 08:43:53.232786 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ddkcm" event={"ID":"431c8c01-259f-4ed7-8244-f16a549042ef","Type":"ContainerDied","Data":"f4774067f075249866c85c6b565ab3c2fd9a387c84e9810accf0a1e09da08a54"} Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.728673 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.835724 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bxm9\" (UniqueName: \"kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9\") pod \"431c8c01-259f-4ed7-8244-f16a549042ef\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.835826 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data\") pod \"431c8c01-259f-4ed7-8244-f16a549042ef\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.835935 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data\") pod \"431c8c01-259f-4ed7-8244-f16a549042ef\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.836128 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle\") pod \"431c8c01-259f-4ed7-8244-f16a549042ef\" (UID: \"431c8c01-259f-4ed7-8244-f16a549042ef\") " Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.846737 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "431c8c01-259f-4ed7-8244-f16a549042ef" (UID: "431c8c01-259f-4ed7-8244-f16a549042ef"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.849167 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9" (OuterVolumeSpecName: "kube-api-access-8bxm9") pod "431c8c01-259f-4ed7-8244-f16a549042ef" (UID: "431c8c01-259f-4ed7-8244-f16a549042ef"). InnerVolumeSpecName "kube-api-access-8bxm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.851232 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data" (OuterVolumeSpecName: "config-data") pod "431c8c01-259f-4ed7-8244-f16a549042ef" (UID: "431c8c01-259f-4ed7-8244-f16a549042ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.876257 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "431c8c01-259f-4ed7-8244-f16a549042ef" (UID: "431c8c01-259f-4ed7-8244-f16a549042ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.938258 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bxm9\" (UniqueName: \"kubernetes.io/projected/431c8c01-259f-4ed7-8244-f16a549042ef-kube-api-access-8bxm9\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.938545 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.938626 5002 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:54 crc kubenswrapper[5002]: I1014 08:43:54.938696 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431c8c01-259f-4ed7-8244-f16a549042ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.256528 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ddkcm" event={"ID":"431c8c01-259f-4ed7-8244-f16a549042ef","Type":"ContainerDied","Data":"eca1f6ef4e7f6e9b5c34548a5b7825c4a62bc28603b16446a712bbff55c668b4"} Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.256582 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eca1f6ef4e7f6e9b5c34548a5b7825c4a62bc28603b16446a712bbff55c668b4" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.256609 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ddkcm" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.610897 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: E1014 08:43:55.611727 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431c8c01-259f-4ed7-8244-f16a549042ef" containerName="manila-db-sync" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.611752 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="431c8c01-259f-4ed7-8244-f16a549042ef" containerName="manila-db-sync" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.612006 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="431c8c01-259f-4ed7-8244-f16a549042ef" containerName="manila-db-sync" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.613708 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.624943 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.633371 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-lcb82" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.633544 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.633827 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.634505 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.636983 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.639672 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.644555 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.662250 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754341 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754414 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754499 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754525 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754549 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754568 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754600 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754631 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jz7g\" (UniqueName: \"kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754657 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754679 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754723 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754742 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754791 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.754821 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld48p\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.775431 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54d78c4b9-jm4l9"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.781153 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.793469 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54d78c4b9-jm4l9"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.851126 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.852544 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.860402 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862154 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-nb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862195 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jz7g\" (UniqueName: \"kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862215 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862464 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862511 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862589 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862613 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862642 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-dns-svc\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862661 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862689 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-sb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862722 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7vlh\" (UniqueName: \"kubernetes.io/projected/244cc8e3-294e-4c7f-b77f-10596f9e56ab-kube-api-access-d7vlh\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862763 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862811 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld48p\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862874 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862908 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.862992 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863031 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863068 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863121 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863154 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-config\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863240 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863288 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863324 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863351 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863370 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863386 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqmg4\" (UniqueName: \"kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.863440 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.864403 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.865195 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.866791 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.866804 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.869965 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.869995 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.870045 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.870208 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.870579 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.871613 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.872264 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.889255 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.891419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld48p\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.896997 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.908880 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jz7g\" (UniqueName: \"kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g\") pod \"manila-scheduler-0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.960663 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970328 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-sb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970369 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7vlh\" (UniqueName: \"kubernetes.io/projected/244cc8e3-294e-4c7f-b77f-10596f9e56ab-kube-api-access-d7vlh\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970411 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970430 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970466 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970500 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970523 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-config\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970561 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970590 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqmg4\" (UniqueName: \"kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970628 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-nb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970653 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970687 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-dns-svc\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.970703 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.974285 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.974964 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-sb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.977023 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.980164 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.980677 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.980899 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.980960 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-ovsdbserver-nb\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.981193 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.981638 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-dns-svc\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.982210 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-config\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.982232 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/244cc8e3-294e-4c7f-b77f-10596f9e56ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.985260 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.997964 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7vlh\" (UniqueName: \"kubernetes.io/projected/244cc8e3-294e-4c7f-b77f-10596f9e56ab-kube-api-access-d7vlh\") pod \"dnsmasq-dns-54d78c4b9-jm4l9\" (UID: \"244cc8e3-294e-4c7f-b77f-10596f9e56ab\") " pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:55 crc kubenswrapper[5002]: I1014 08:43:55.999198 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqmg4\" (UniqueName: \"kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4\") pod \"manila-api-0\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " pod="openstack/manila-api-0" Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.105254 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.277528 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.562255 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:43:56 crc kubenswrapper[5002]: W1014 08:43:56.566585 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a145b9e_3b5d_4307_a6a6_f8c0ce3afab8.slice/crio-fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27 WatchSource:0}: Error finding container fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27: Status 404 returned error can't find the container with id fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27 Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.587140 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.657522 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54d78c4b9-jm4l9"] Oct 14 08:43:56 crc kubenswrapper[5002]: W1014 08:43:56.657782 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod244cc8e3_294e_4c7f_b77f_10596f9e56ab.slice/crio-1f1d8cb2a404775d3ac10a3baae2e78900018a3776eb6537635a2163fb86807e WatchSource:0}: Error finding container 1f1d8cb2a404775d3ac10a3baae2e78900018a3776eb6537635a2163fb86807e: Status 404 returned error can't find the container with id 1f1d8cb2a404775d3ac10a3baae2e78900018a3776eb6537635a2163fb86807e Oct 14 08:43:56 crc kubenswrapper[5002]: I1014 08:43:56.875784 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:43:56 crc kubenswrapper[5002]: W1014 08:43:56.888992 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfc0b2b8_0d19_47bf_8687_7ec9200344e8.slice/crio-511e78e05f277059f1000b9d27158a7ea435fdb0907113dcea26d0af6ab2c426 WatchSource:0}: Error finding container 511e78e05f277059f1000b9d27158a7ea435fdb0907113dcea26d0af6ab2c426: Status 404 returned error can't find the container with id 511e78e05f277059f1000b9d27158a7ea435fdb0907113dcea26d0af6ab2c426 Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.276947 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerStarted","Data":"91bde49da7eebae6ee7621634c7a6a769811252ccf2a55a43039b4afffd47a79"} Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.282701 5002 generic.go:334] "Generic (PLEG): container finished" podID="244cc8e3-294e-4c7f-b77f-10596f9e56ab" containerID="d1e0365d2da04d56ddd43d09d3ea7ceb752d3825fbb01b2b3d4fbee686ed3507" exitCode=0 Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.282734 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" event={"ID":"244cc8e3-294e-4c7f-b77f-10596f9e56ab","Type":"ContainerDied","Data":"d1e0365d2da04d56ddd43d09d3ea7ceb752d3825fbb01b2b3d4fbee686ed3507"} Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.282776 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" event={"ID":"244cc8e3-294e-4c7f-b77f-10596f9e56ab","Type":"ContainerStarted","Data":"1f1d8cb2a404775d3ac10a3baae2e78900018a3776eb6537635a2163fb86807e"} Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.284099 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerStarted","Data":"fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27"} Oct 14 08:43:57 crc kubenswrapper[5002]: I1014 08:43:57.289355 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerStarted","Data":"511e78e05f277059f1000b9d27158a7ea435fdb0907113dcea26d0af6ab2c426"} Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.323150 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerStarted","Data":"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5"} Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.323572 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerStarted","Data":"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf"} Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.323923 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.361356 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" event={"ID":"244cc8e3-294e-4c7f-b77f-10596f9e56ab","Type":"ContainerStarted","Data":"9316886b911d92757ec8b157665e9309e74a629a81eeacdaf2566a3fee32599d"} Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.361558 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.409335 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.409318162 podStartE2EDuration="3.409318162s" podCreationTimestamp="2025-10-14 08:43:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:43:58.356828815 +0000 UTC m=+3171.338068277" watchObservedRunningTime="2025-10-14 08:43:58.409318162 +0000 UTC m=+3171.390557604" Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.410570 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" podStartSLOduration=3.410563375 podStartE2EDuration="3.410563375s" podCreationTimestamp="2025-10-14 08:43:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:43:58.38646394 +0000 UTC m=+3171.367703412" watchObservedRunningTime="2025-10-14 08:43:58.410563375 +0000 UTC m=+3171.391802827" Oct 14 08:43:58 crc kubenswrapper[5002]: I1014 08:43:58.999386 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:43:59 crc kubenswrapper[5002]: I1014 08:43:59.379819 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerStarted","Data":"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28"} Oct 14 08:43:59 crc kubenswrapper[5002]: I1014 08:43:59.380142 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerStarted","Data":"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437"} Oct 14 08:43:59 crc kubenswrapper[5002]: I1014 08:43:59.404896 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.178068005 podStartE2EDuration="4.40486894s" podCreationTimestamp="2025-10-14 08:43:55 +0000 UTC" firstStartedPulling="2025-10-14 08:43:56.60010717 +0000 UTC m=+3169.581346622" lastFinishedPulling="2025-10-14 08:43:57.826908095 +0000 UTC m=+3170.808147557" observedRunningTime="2025-10-14 08:43:59.397087451 +0000 UTC m=+3172.378326913" watchObservedRunningTime="2025-10-14 08:43:59.40486894 +0000 UTC m=+3172.386108392" Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.319951 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.320207 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-central-agent" containerID="cri-o://bbef91002f42dce8d3ae557c3f9842e0589749f929357b4acdb773f81c4373a0" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.320547 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="sg-core" containerID="cri-o://6d30cf1060a60054748dff0eb336d3fbebc717dccd47808ca981fe76d3a77992" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.320527 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="proxy-httpd" containerID="cri-o://53c4ee8860f08e35c23df2d3535b4380e7996a01e9b4483f0fb50f76f349a3f5" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.320600 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-notification-agent" containerID="cri-o://beabcf21650b91cde42a85e773d3aaa2388b081e7251ca8d8d6fc0d3ac407829" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.387167 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api-log" containerID="cri-o://0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.387198 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api" containerID="cri-o://358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" gracePeriod=30 Oct 14 08:44:00 crc kubenswrapper[5002]: I1014 08:44:00.904496 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.116726 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117043 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117089 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117132 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqmg4\" (UniqueName: \"kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117239 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117317 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.117429 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle\") pod \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\" (UID: \"cfc0b2b8-0d19-47bf-8687-7ec9200344e8\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.122272 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.122450 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs" (OuterVolumeSpecName: "logs") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.122950 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.122988 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4" (OuterVolumeSpecName: "kube-api-access-kqmg4") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "kube-api-access-kqmg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.125969 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts" (OuterVolumeSpecName: "scripts") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.148558 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.171734 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data" (OuterVolumeSpecName: "config-data") pod "cfc0b2b8-0d19-47bf-8687-7ec9200344e8" (UID: "cfc0b2b8-0d19-47bf-8687-7ec9200344e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219109 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219154 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219168 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219182 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqmg4\" (UniqueName: \"kubernetes.io/projected/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-kube-api-access-kqmg4\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219196 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219207 5002 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.219219 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc0b2b8-0d19-47bf-8687-7ec9200344e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396427 5002 generic.go:334] "Generic (PLEG): container finished" podID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerID="358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" exitCode=0 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396721 5002 generic.go:334] "Generic (PLEG): container finished" podID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerID="0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" exitCode=143 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396593 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerDied","Data":"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396689 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396790 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerDied","Data":"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396808 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cfc0b2b8-0d19-47bf-8687-7ec9200344e8","Type":"ContainerDied","Data":"511e78e05f277059f1000b9d27158a7ea435fdb0907113dcea26d0af6ab2c426"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.396825 5002 scope.go:117] "RemoveContainer" containerID="358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408702 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerID="53c4ee8860f08e35c23df2d3535b4380e7996a01e9b4483f0fb50f76f349a3f5" exitCode=0 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408732 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerID="6d30cf1060a60054748dff0eb336d3fbebc717dccd47808ca981fe76d3a77992" exitCode=2 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408739 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerID="beabcf21650b91cde42a85e773d3aaa2388b081e7251ca8d8d6fc0d3ac407829" exitCode=0 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408751 5002 generic.go:334] "Generic (PLEG): container finished" podID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerID="bbef91002f42dce8d3ae557c3f9842e0589749f929357b4acdb773f81c4373a0" exitCode=0 Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408770 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerDied","Data":"53c4ee8860f08e35c23df2d3535b4380e7996a01e9b4483f0fb50f76f349a3f5"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408794 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerDied","Data":"6d30cf1060a60054748dff0eb336d3fbebc717dccd47808ca981fe76d3a77992"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408804 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerDied","Data":"beabcf21650b91cde42a85e773d3aaa2388b081e7251ca8d8d6fc0d3ac407829"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.408813 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerDied","Data":"bbef91002f42dce8d3ae557c3f9842e0589749f929357b4acdb773f81c4373a0"} Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.426370 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.429068 5002 scope.go:117] "RemoveContainer" containerID="0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.434241 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.449369 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 14 08:44:01 crc kubenswrapper[5002]: E1014 08:44:01.449786 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.449803 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api" Oct 14 08:44:01 crc kubenswrapper[5002]: E1014 08:44:01.449829 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api-log" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.449848 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api-log" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.450035 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api-log" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.450052 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" containerName="manila-api" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.454579 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.461830 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.462914 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.463009 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.463409 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.542810 5002 scope.go:117] "RemoveContainer" containerID="358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" Oct 14 08:44:01 crc kubenswrapper[5002]: E1014 08:44:01.548584 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5\": container with ID starting with 358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5 not found: ID does not exist" containerID="358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.548633 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5"} err="failed to get container status \"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5\": rpc error: code = NotFound desc = could not find container \"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5\": container with ID starting with 358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5 not found: ID does not exist" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.548662 5002 scope.go:117] "RemoveContainer" containerID="0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" Oct 14 08:44:01 crc kubenswrapper[5002]: E1014 08:44:01.549213 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf\": container with ID starting with 0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf not found: ID does not exist" containerID="0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.549248 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf"} err="failed to get container status \"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf\": rpc error: code = NotFound desc = could not find container \"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf\": container with ID starting with 0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf not found: ID does not exist" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.549272 5002 scope.go:117] "RemoveContainer" containerID="358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.549608 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5"} err="failed to get container status \"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5\": rpc error: code = NotFound desc = could not find container \"358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5\": container with ID starting with 358ad92fa85e98a37750f09097144bcd4b8b3b75a881f3e620f3cd2a34bb14e5 not found: ID does not exist" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.549629 5002 scope.go:117] "RemoveContainer" containerID="0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.549903 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf"} err="failed to get container status \"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf\": rpc error: code = NotFound desc = could not find container \"0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf\": container with ID starting with 0b6e601ccf576332082af874999cf8fa863f3b1440f57f25a5cded18195833cf not found: ID does not exist" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.626435 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-scripts\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.626477 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978c2b52-ead1-4ca3-924e-83ed4bb321a5-logs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.626509 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627395 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627476 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctz4v\" (UniqueName: \"kubernetes.io/projected/978c2b52-ead1-4ca3-924e-83ed4bb321a5-kube-api-access-ctz4v\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627532 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/978c2b52-ead1-4ca3-924e-83ed4bb321a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627584 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627829 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.627977 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data-custom\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730393 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data-custom\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730506 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-scripts\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730525 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978c2b52-ead1-4ca3-924e-83ed4bb321a5-logs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730549 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730609 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.730644 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctz4v\" (UniqueName: \"kubernetes.io/projected/978c2b52-ead1-4ca3-924e-83ed4bb321a5-kube-api-access-ctz4v\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.731577 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/978c2b52-ead1-4ca3-924e-83ed4bb321a5-logs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.731577 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc0b2b8-0d19-47bf-8687-7ec9200344e8" path="/var/lib/kubelet/pods/cfc0b2b8-0d19-47bf-8687-7ec9200344e8/volumes" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.733282 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/978c2b52-ead1-4ca3-924e-83ed4bb321a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.733326 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.733449 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/978c2b52-ead1-4ca3-924e-83ed4bb321a5-etc-machine-id\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.733706 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.735814 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.736300 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-internal-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.736340 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-config-data-custom\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.737023 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-public-tls-certs\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.739319 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.741310 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/978c2b52-ead1-4ca3-924e-83ed4bb321a5-scripts\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.750224 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctz4v\" (UniqueName: \"kubernetes.io/projected/978c2b52-ead1-4ca3-924e-83ed4bb321a5-kube-api-access-ctz4v\") pod \"manila-api-0\" (UID: \"978c2b52-ead1-4ca3-924e-83ed4bb321a5\") " pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.812682 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.816161 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.834529 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.834796 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvcwf\" (UniqueName: \"kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.834898 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.834984 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.835077 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.835159 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.835226 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.835312 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml\") pod \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\" (UID: \"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a\") " Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.836433 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.837256 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.838808 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.838830 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.867050 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf" (OuterVolumeSpecName: "kube-api-access-wvcwf") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "kube-api-access-wvcwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.871085 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts" (OuterVolumeSpecName: "scripts") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.920944 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.922283 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.939563 5002 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.939594 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.939605 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:01 crc kubenswrapper[5002]: I1014 08:44:01.939613 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvcwf\" (UniqueName: \"kubernetes.io/projected/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-kube-api-access-wvcwf\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.003401 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data" (OuterVolumeSpecName: "config-data") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.007110 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" (UID: "c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.043469 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.043498 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.423066 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a","Type":"ContainerDied","Data":"11531cdd2629e2633cb4e436d819a692a48d5c590d0f65370f9431905f3fb52a"} Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.423149 5002 scope.go:117] "RemoveContainer" containerID="53c4ee8860f08e35c23df2d3535b4380e7996a01e9b4483f0fb50f76f349a3f5" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.423107 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.543821 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.553619 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.569553 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:02 crc kubenswrapper[5002]: E1014 08:44:02.570092 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="proxy-httpd" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570117 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="proxy-httpd" Oct 14 08:44:02 crc kubenswrapper[5002]: E1014 08:44:02.570151 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-notification-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570162 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-notification-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: E1014 08:44:02.570180 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="sg-core" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570189 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="sg-core" Oct 14 08:44:02 crc kubenswrapper[5002]: E1014 08:44:02.570211 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-central-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570220 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-central-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570556 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="proxy-httpd" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570586 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-central-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570605 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="ceilometer-notification-agent" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.570614 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" containerName="sg-core" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.572757 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.578403 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.578746 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.578928 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.587324 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.658947 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659209 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659514 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96nm\" (UniqueName: \"kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659604 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659676 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659756 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.659992 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762269 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96nm\" (UniqueName: \"kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762346 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762382 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762441 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762506 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762533 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762602 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.762647 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.763235 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.763875 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.768770 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.768797 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.769327 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.772141 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.772735 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.778464 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96nm\" (UniqueName: \"kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm\") pod \"ceilometer-0\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " pod="openstack/ceilometer-0" Oct 14 08:44:02 crc kubenswrapper[5002]: I1014 08:44:02.902076 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:03 crc kubenswrapper[5002]: I1014 08:44:03.733871 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a" path="/var/lib/kubelet/pods/c5a8ceb1-0d12-4bdb-b903-b99c044d8a4a/volumes" Oct 14 08:44:04 crc kubenswrapper[5002]: I1014 08:44:04.435113 5002 scope.go:117] "RemoveContainer" containerID="6d30cf1060a60054748dff0eb336d3fbebc717dccd47808ca981fe76d3a77992" Oct 14 08:44:04 crc kubenswrapper[5002]: I1014 08:44:04.586662 5002 scope.go:117] "RemoveContainer" containerID="beabcf21650b91cde42a85e773d3aaa2388b081e7251ca8d8d6fc0d3ac407829" Oct 14 08:44:04 crc kubenswrapper[5002]: I1014 08:44:04.743150 5002 scope.go:117] "RemoveContainer" containerID="bbef91002f42dce8d3ae557c3f9842e0589749f929357b4acdb773f81c4373a0" Oct 14 08:44:05 crc kubenswrapper[5002]: I1014 08:44:05.158100 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:05 crc kubenswrapper[5002]: W1014 08:44:05.162975 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ed8447d_286d_4367_adb4_697e11493e60.slice/crio-c15363787398ffdf5643fcb547245539a3fd6c3cfe5a1a0f2d562f5aa5e6710e WatchSource:0}: Error finding container c15363787398ffdf5643fcb547245539a3fd6c3cfe5a1a0f2d562f5aa5e6710e: Status 404 returned error can't find the container with id c15363787398ffdf5643fcb547245539a3fd6c3cfe5a1a0f2d562f5aa5e6710e Oct 14 08:44:05 crc kubenswrapper[5002]: I1014 08:44:05.175117 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 08:44:05 crc kubenswrapper[5002]: W1014 08:44:05.178144 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978c2b52_ead1_4ca3_924e_83ed4bb321a5.slice/crio-d4fb9cf7a3eb11a0dbec5efb7fcfbd8791ddcd84115b67f060929e81b83676d9 WatchSource:0}: Error finding container d4fb9cf7a3eb11a0dbec5efb7fcfbd8791ddcd84115b67f060929e81b83676d9: Status 404 returned error can't find the container with id d4fb9cf7a3eb11a0dbec5efb7fcfbd8791ddcd84115b67f060929e81b83676d9 Oct 14 08:44:05 crc kubenswrapper[5002]: I1014 08:44:05.461745 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerStarted","Data":"c15363787398ffdf5643fcb547245539a3fd6c3cfe5a1a0f2d562f5aa5e6710e"} Oct 14 08:44:05 crc kubenswrapper[5002]: I1014 08:44:05.464549 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"978c2b52-ead1-4ca3-924e-83ed4bb321a5","Type":"ContainerStarted","Data":"d4fb9cf7a3eb11a0dbec5efb7fcfbd8791ddcd84115b67f060929e81b83676d9"} Oct 14 08:44:05 crc kubenswrapper[5002]: I1014 08:44:05.986224 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.108363 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54d78c4b9-jm4l9" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.184154 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.185265 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="dnsmasq-dns" containerID="cri-o://80769c05d7af0588e89ecda6c65d5f674afd365655adfc5a296051a624518034" gracePeriod=10 Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.492396 5002 generic.go:334] "Generic (PLEG): container finished" podID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerID="80769c05d7af0588e89ecda6c65d5f674afd365655adfc5a296051a624518034" exitCode=0 Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.492479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" event={"ID":"49cf58a5-bf2a-46bb-b08f-94e85686c2a6","Type":"ContainerDied","Data":"80769c05d7af0588e89ecda6c65d5f674afd365655adfc5a296051a624518034"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.504128 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerStarted","Data":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.513433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"978c2b52-ead1-4ca3-924e-83ed4bb321a5","Type":"ContainerStarted","Data":"3cd9c6a523fe82ad56c119151577a6360e70d7c4030180bfe194814808570c2c"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.513485 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"978c2b52-ead1-4ca3-924e-83ed4bb321a5","Type":"ContainerStarted","Data":"91bf2b12a768655e74954f125944e3a4b74f4510798a9272c29200f276305401"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.513524 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.527372 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerStarted","Data":"71cec4a0ac5eb7cd96f7f217f31c7f62dafa0bf16492820ad7af2e3ba551d454"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.527689 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerStarted","Data":"a4e1b4f134a570a93b23a42a88a8016a0853080e9e2c12779e27ff3181e43e2b"} Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.540292 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.54026948 podStartE2EDuration="5.54026948s" podCreationTimestamp="2025-10-14 08:44:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:44:06.529030499 +0000 UTC m=+3179.510269971" watchObservedRunningTime="2025-10-14 08:44:06.54026948 +0000 UTC m=+3179.521508932" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.594789 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.5638769420000003 podStartE2EDuration="11.59476674s" podCreationTimestamp="2025-10-14 08:43:55 +0000 UTC" firstStartedPulling="2025-10-14 08:43:56.568620605 +0000 UTC m=+3169.549860047" lastFinishedPulling="2025-10-14 08:44:04.599510353 +0000 UTC m=+3177.580749845" observedRunningTime="2025-10-14 08:44:06.569558235 +0000 UTC m=+3179.550797707" watchObservedRunningTime="2025-10-14 08:44:06.59476674 +0000 UTC m=+3179.576006192" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.715975 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.777715 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.843441 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.879524 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgfpd\" (UniqueName: \"kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.879595 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.879628 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.879685 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.879793 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam\") pod \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\" (UID: \"49cf58a5-bf2a-46bb-b08f-94e85686c2a6\") " Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.880494 5002 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.884030 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd" (OuterVolumeSpecName: "kube-api-access-wgfpd") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "kube-api-access-wgfpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.936258 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.940417 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.947077 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config" (OuterVolumeSpecName: "config") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.949391 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "49cf58a5-bf2a-46bb-b08f-94e85686c2a6" (UID: "49cf58a5-bf2a-46bb-b08f-94e85686c2a6"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.981586 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgfpd\" (UniqueName: \"kubernetes.io/projected/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-kube-api-access-wgfpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.981620 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.981632 5002 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.981643 5002 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:06 crc kubenswrapper[5002]: I1014 08:44:06.981651 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/49cf58a5-bf2a-46bb-b08f-94e85686c2a6-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.539141 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" event={"ID":"49cf58a5-bf2a-46bb-b08f-94e85686c2a6","Type":"ContainerDied","Data":"f11c8cb3ee322722e199bbbf922869636a44157c7948e6c3f59663f5d44978c1"} Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.539676 5002 scope.go:117] "RemoveContainer" containerID="80769c05d7af0588e89ecda6c65d5f674afd365655adfc5a296051a624518034" Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.539175 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c8c578c57-x8rz8" Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.543380 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerStarted","Data":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.626227 5002 scope.go:117] "RemoveContainer" containerID="ae26742f397f08dbff20c2d52ada7f2514ff2daf722a6ae5d74e34105097687b" Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.631345 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.650607 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c8c578c57-x8rz8"] Oct 14 08:44:07 crc kubenswrapper[5002]: I1014 08:44:07.754686 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" path="/var/lib/kubelet/pods/49cf58a5-bf2a-46bb-b08f-94e85686c2a6/volumes" Oct 14 08:44:08 crc kubenswrapper[5002]: I1014 08:44:08.555314 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerStarted","Data":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} Oct 14 08:44:09 crc kubenswrapper[5002]: I1014 08:44:09.341904 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.584351 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerStarted","Data":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.584694 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-central-agent" containerID="cri-o://5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" gracePeriod=30 Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.584953 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="proxy-httpd" containerID="cri-o://1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" gracePeriod=30 Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.585068 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="sg-core" containerID="cri-o://4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" gracePeriod=30 Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.585166 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-notification-agent" containerID="cri-o://99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" gracePeriod=30 Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.584720 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:44:10 crc kubenswrapper[5002]: I1014 08:44:10.608261 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.181875907 podStartE2EDuration="8.608242522s" podCreationTimestamp="2025-10-14 08:44:02 +0000 UTC" firstStartedPulling="2025-10-14 08:44:05.166336303 +0000 UTC m=+3178.147575795" lastFinishedPulling="2025-10-14 08:44:09.592702958 +0000 UTC m=+3182.573942410" observedRunningTime="2025-10-14 08:44:10.605333513 +0000 UTC m=+3183.586572965" watchObservedRunningTime="2025-10-14 08:44:10.608242522 +0000 UTC m=+3183.589481974" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.348751 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.542117 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.542392 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543402 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543585 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543715 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543807 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543926 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96nm\" (UniqueName: \"kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.543990 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.544017 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs\") pod \"6ed8447d-286d-4367-adb4-697e11493e60\" (UID: \"6ed8447d-286d-4367-adb4-697e11493e60\") " Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.544488 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.544817 5002 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.544860 5002 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ed8447d-286d-4367-adb4-697e11493e60-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.548249 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm" (OuterVolumeSpecName: "kube-api-access-z96nm") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "kube-api-access-z96nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.549041 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts" (OuterVolumeSpecName: "scripts") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.574079 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596563 5002 generic.go:334] "Generic (PLEG): container finished" podID="6ed8447d-286d-4367-adb4-697e11493e60" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" exitCode=0 Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596594 5002 generic.go:334] "Generic (PLEG): container finished" podID="6ed8447d-286d-4367-adb4-697e11493e60" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" exitCode=2 Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596602 5002 generic.go:334] "Generic (PLEG): container finished" podID="6ed8447d-286d-4367-adb4-697e11493e60" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" exitCode=0 Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596609 5002 generic.go:334] "Generic (PLEG): container finished" podID="6ed8447d-286d-4367-adb4-697e11493e60" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" exitCode=0 Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596620 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerDied","Data":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596681 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerDied","Data":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596695 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerDied","Data":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596704 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerDied","Data":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596714 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ed8447d-286d-4367-adb4-697e11493e60","Type":"ContainerDied","Data":"c15363787398ffdf5643fcb547245539a3fd6c3cfe5a1a0f2d562f5aa5e6710e"} Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.596734 5002 scope.go:117] "RemoveContainer" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.597985 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.603444 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.627120 5002 scope.go:117] "RemoveContainer" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.628633 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.646943 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.646993 5002 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.647011 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96nm\" (UniqueName: \"kubernetes.io/projected/6ed8447d-286d-4367-adb4-697e11493e60-kube-api-access-z96nm\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.647030 5002 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.647049 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.647992 5002 scope.go:117] "RemoveContainer" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.669378 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data" (OuterVolumeSpecName: "config-data") pod "6ed8447d-286d-4367-adb4-697e11493e60" (UID: "6ed8447d-286d-4367-adb4-697e11493e60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.672964 5002 scope.go:117] "RemoveContainer" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.694545 5002 scope.go:117] "RemoveContainer" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.695087 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": container with ID starting with 1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c not found: ID does not exist" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695110 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} err="failed to get container status \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": rpc error: code = NotFound desc = could not find container \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": container with ID starting with 1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695130 5002 scope.go:117] "RemoveContainer" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.695434 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": container with ID starting with 4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333 not found: ID does not exist" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695455 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} err="failed to get container status \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": rpc error: code = NotFound desc = could not find container \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": container with ID starting with 4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695467 5002 scope.go:117] "RemoveContainer" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.695695 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": container with ID starting with 99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990 not found: ID does not exist" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695712 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} err="failed to get container status \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": rpc error: code = NotFound desc = could not find container \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": container with ID starting with 99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.695725 5002 scope.go:117] "RemoveContainer" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.696021 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": container with ID starting with 5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa not found: ID does not exist" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696037 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} err="failed to get container status \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": rpc error: code = NotFound desc = could not find container \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": container with ID starting with 5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696048 5002 scope.go:117] "RemoveContainer" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696538 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} err="failed to get container status \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": rpc error: code = NotFound desc = could not find container \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": container with ID starting with 1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696553 5002 scope.go:117] "RemoveContainer" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696774 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} err="failed to get container status \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": rpc error: code = NotFound desc = could not find container \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": container with ID starting with 4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.696786 5002 scope.go:117] "RemoveContainer" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697138 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} err="failed to get container status \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": rpc error: code = NotFound desc = could not find container \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": container with ID starting with 99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697152 5002 scope.go:117] "RemoveContainer" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697427 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} err="failed to get container status \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": rpc error: code = NotFound desc = could not find container \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": container with ID starting with 5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697439 5002 scope.go:117] "RemoveContainer" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697695 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} err="failed to get container status \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": rpc error: code = NotFound desc = could not find container \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": container with ID starting with 1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697706 5002 scope.go:117] "RemoveContainer" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697910 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} err="failed to get container status \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": rpc error: code = NotFound desc = could not find container \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": container with ID starting with 4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.697922 5002 scope.go:117] "RemoveContainer" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698065 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} err="failed to get container status \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": rpc error: code = NotFound desc = could not find container \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": container with ID starting with 99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698076 5002 scope.go:117] "RemoveContainer" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698223 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} err="failed to get container status \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": rpc error: code = NotFound desc = could not find container \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": container with ID starting with 5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698235 5002 scope.go:117] "RemoveContainer" containerID="1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698354 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c"} err="failed to get container status \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": rpc error: code = NotFound desc = could not find container \"1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c\": container with ID starting with 1852249e3681381e9687bfabda631c03d59db6253dadc26db3a00b1779e9764c not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698365 5002 scope.go:117] "RemoveContainer" containerID="4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698636 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333"} err="failed to get container status \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": rpc error: code = NotFound desc = could not find container \"4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333\": container with ID starting with 4b88f48f6b773fb2863c559116a61756ede1fc3fcee207ce685a304e062a6333 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698648 5002 scope.go:117] "RemoveContainer" containerID="99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698792 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990"} err="failed to get container status \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": rpc error: code = NotFound desc = could not find container \"99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990\": container with ID starting with 99b0f9d324d63a8a27b5f4a62fd48de9670190de4f3152716a5cfb260dca1990 not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.698803 5002 scope.go:117] "RemoveContainer" containerID="5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.699038 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa"} err="failed to get container status \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": rpc error: code = NotFound desc = could not find container \"5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa\": container with ID starting with 5b143f92859fcc3326fa2a6b4b17412ed78de293cdf92cd8c10b99b8648f4efa not found: ID does not exist" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.748875 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ed8447d-286d-4367-adb4-697e11493e60-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.948406 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.974823 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991146 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991705 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="dnsmasq-dns" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991737 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="dnsmasq-dns" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991756 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="sg-core" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991767 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="sg-core" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991794 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="init" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991806 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="init" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991827 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="proxy-httpd" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991856 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="proxy-httpd" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991871 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-central-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991882 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-central-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: E1014 08:44:11.991904 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-notification-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.991917 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-notification-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.992201 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="sg-core" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.992235 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-central-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.992257 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="49cf58a5-bf2a-46bb-b08f-94e85686c2a6" containerName="dnsmasq-dns" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.992267 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="ceilometer-notification-agent" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.992284 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed8447d-286d-4367-adb4-697e11493e60" containerName="proxy-httpd" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.994926 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:11 crc kubenswrapper[5002]: I1014 08:44:11.998078 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.000144 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.000278 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.002408 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.056930 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-config-data\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.057229 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.057527 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-run-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.057747 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.057919 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqtx9\" (UniqueName: \"kubernetes.io/projected/41dfcd37-c0b9-426a-bbf9-fca697e2368a-kube-api-access-qqtx9\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.058152 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-log-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.058350 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.058538 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-scripts\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160082 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-run-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160383 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160485 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqtx9\" (UniqueName: \"kubernetes.io/projected/41dfcd37-c0b9-426a-bbf9-fca697e2368a-kube-api-access-qqtx9\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160602 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-log-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160732 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.160881 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-scripts\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.161023 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-config-data\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.161142 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.162248 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-run-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.162700 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dfcd37-c0b9-426a-bbf9-fca697e2368a-log-httpd\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.165237 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.166374 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-scripts\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.167052 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-config-data\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.168408 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.177342 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dfcd37-c0b9-426a-bbf9-fca697e2368a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.179170 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqtx9\" (UniqueName: \"kubernetes.io/projected/41dfcd37-c0b9-426a-bbf9-fca697e2368a-kube-api-access-qqtx9\") pod \"ceilometer-0\" (UID: \"41dfcd37-c0b9-426a-bbf9-fca697e2368a\") " pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.328532 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 08:44:12 crc kubenswrapper[5002]: I1014 08:44:12.853813 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 08:44:12 crc kubenswrapper[5002]: W1014 08:44:12.866553 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41dfcd37_c0b9_426a_bbf9_fca697e2368a.slice/crio-e09531e3ae51398a6bfe1792b37a4e9e394ff88099253128c810b67c900176bb WatchSource:0}: Error finding container e09531e3ae51398a6bfe1792b37a4e9e394ff88099253128c810b67c900176bb: Status 404 returned error can't find the container with id e09531e3ae51398a6bfe1792b37a4e9e394ff88099253128c810b67c900176bb Oct 14 08:44:13 crc kubenswrapper[5002]: I1014 08:44:13.617318 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dfcd37-c0b9-426a-bbf9-fca697e2368a","Type":"ContainerStarted","Data":"e09531e3ae51398a6bfe1792b37a4e9e394ff88099253128c810b67c900176bb"} Oct 14 08:44:13 crc kubenswrapper[5002]: I1014 08:44:13.736146 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed8447d-286d-4367-adb4-697e11493e60" path="/var/lib/kubelet/pods/6ed8447d-286d-4367-adb4-697e11493e60/volumes" Oct 14 08:44:14 crc kubenswrapper[5002]: I1014 08:44:14.629583 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dfcd37-c0b9-426a-bbf9-fca697e2368a","Type":"ContainerStarted","Data":"ed2473fefb17505683d270515c94fa85e659aea4db33d5e1fb4472b57d891919"} Oct 14 08:44:15 crc kubenswrapper[5002]: I1014 08:44:15.643119 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dfcd37-c0b9-426a-bbf9-fca697e2368a","Type":"ContainerStarted","Data":"f18a3efe3604103db6831c1afa88da511ae5060dc31255dafb2011ab76f8ffd3"} Oct 14 08:44:15 crc kubenswrapper[5002]: I1014 08:44:15.961866 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 14 08:44:16 crc kubenswrapper[5002]: I1014 08:44:16.659034 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dfcd37-c0b9-426a-bbf9-fca697e2368a","Type":"ContainerStarted","Data":"5476c6f333bd8571734cedd217e30bc746d720754e7de53429c0fc5fb0f919f1"} Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.485217 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.548071 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.566348 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.614548 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.676688 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="manila-scheduler" containerID="cri-o://c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437" gracePeriod=30 Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.677540 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dfcd37-c0b9-426a-bbf9-fca697e2368a","Type":"ContainerStarted","Data":"121516a2dc47d3caaf6f9b12092d719754c903f734f97b7b12009f3872ff72a3"} Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.678095 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="probe" containerID="cri-o://96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28" gracePeriod=30 Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.678387 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.679090 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="manila-share" containerID="cri-o://a4e1b4f134a570a93b23a42a88a8016a0853080e9e2c12779e27ff3181e43e2b" gracePeriod=30 Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.679256 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="probe" containerID="cri-o://71cec4a0ac5eb7cd96f7f217f31c7f62dafa0bf16492820ad7af2e3ba551d454" gracePeriod=30 Oct 14 08:44:17 crc kubenswrapper[5002]: I1014 08:44:17.708068 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.597094126 podStartE2EDuration="6.708049509s" podCreationTimestamp="2025-10-14 08:44:11 +0000 UTC" firstStartedPulling="2025-10-14 08:44:12.874082781 +0000 UTC m=+3185.855322233" lastFinishedPulling="2025-10-14 08:44:16.985038164 +0000 UTC m=+3189.966277616" observedRunningTime="2025-10-14 08:44:17.706903339 +0000 UTC m=+3190.688142801" watchObservedRunningTime="2025-10-14 08:44:17.708049509 +0000 UTC m=+3190.689288961" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.686855 5002 generic.go:334] "Generic (PLEG): container finished" podID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerID="96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28" exitCode=0 Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.686919 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerDied","Data":"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28"} Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688716 5002 generic.go:334] "Generic (PLEG): container finished" podID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerID="71cec4a0ac5eb7cd96f7f217f31c7f62dafa0bf16492820ad7af2e3ba551d454" exitCode=0 Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688738 5002 generic.go:334] "Generic (PLEG): container finished" podID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerID="a4e1b4f134a570a93b23a42a88a8016a0853080e9e2c12779e27ff3181e43e2b" exitCode=1 Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688743 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerDied","Data":"71cec4a0ac5eb7cd96f7f217f31c7f62dafa0bf16492820ad7af2e3ba551d454"} Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688768 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerDied","Data":"a4e1b4f134a570a93b23a42a88a8016a0853080e9e2c12779e27ff3181e43e2b"} Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688778 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8","Type":"ContainerDied","Data":"fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27"} Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.688786 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fab834313d2a9a93f7ffd7120d602baec8fa1431a50c9a2e2774d5dd42564b27" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.725690 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.903705 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904040 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904188 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904310 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld48p\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904522 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904667 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904751 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904867 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data\") pod \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\" (UID: \"1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8\") " Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.904095 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.905453 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.907530 5002 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.907565 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.910463 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p" (OuterVolumeSpecName: "kube-api-access-ld48p") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "kube-api-access-ld48p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.910640 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts" (OuterVolumeSpecName: "scripts") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.915050 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph" (OuterVolumeSpecName: "ceph") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.917018 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:18 crc kubenswrapper[5002]: I1014 08:44:18.956195 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.009143 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.009182 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.009197 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld48p\" (UniqueName: \"kubernetes.io/projected/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-kube-api-access-ld48p\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.009214 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.009227 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.034134 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data" (OuterVolumeSpecName: "config-data") pod "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" (UID: "1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.111596 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.698244 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.760741 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.783815 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.800431 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:19 crc kubenswrapper[5002]: E1014 08:44:19.801531 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="probe" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.801725 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="probe" Oct 14 08:44:19 crc kubenswrapper[5002]: E1014 08:44:19.801972 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="manila-share" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.802198 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="manila-share" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.802890 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="manila-share" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.803115 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" containerName="probe" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.805742 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.808102 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.812380 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834578 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-ceph\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834629 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834681 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-scripts\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834702 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834861 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834901 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.834962 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tjzz\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-kube-api-access-5tjzz\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936580 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-ceph\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936634 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936681 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-scripts\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936705 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936741 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936771 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936802 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.936818 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tjzz\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-kube-api-access-5tjzz\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.937335 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.937524 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d10fce14-aa5a-466f-948a-7787f2aa8a76-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.941937 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.942105 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.942452 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-scripts\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.955674 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-ceph\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.956027 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fce14-aa5a-466f-948a-7787f2aa8a76-config-data\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:19 crc kubenswrapper[5002]: I1014 08:44:19.961007 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tjzz\" (UniqueName: \"kubernetes.io/projected/d10fce14-aa5a-466f-948a-7787f2aa8a76-kube-api-access-5tjzz\") pod \"manila-share-share1-0\" (UID: \"d10fce14-aa5a-466f-948a-7787f2aa8a76\") " pod="openstack/manila-share-share1-0" Oct 14 08:44:20 crc kubenswrapper[5002]: I1014 08:44:20.137094 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 08:44:20 crc kubenswrapper[5002]: I1014 08:44:20.669247 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 08:44:20 crc kubenswrapper[5002]: I1014 08:44:20.738592 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d10fce14-aa5a-466f-948a-7787f2aa8a76","Type":"ContainerStarted","Data":"dcf742003ed60c33c515ae5eef56e62b32b6c4709463e97fa13ec0c8996fbcea"} Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.369958 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468032 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468075 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jz7g\" (UniqueName: \"kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468120 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468228 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468272 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468313 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id\") pod \"16ab80a6-c4a1-4df0-af42-6711e90313c0\" (UID: \"16ab80a6-c4a1-4df0-af42-6711e90313c0\") " Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.468677 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.469046 5002 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16ab80a6-c4a1-4df0-af42-6711e90313c0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.473029 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g" (OuterVolumeSpecName: "kube-api-access-4jz7g") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "kube-api-access-4jz7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.473137 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts" (OuterVolumeSpecName: "scripts") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.473603 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.515620 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.568952 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data" (OuterVolumeSpecName: "config-data") pod "16ab80a6-c4a1-4df0-af42-6711e90313c0" (UID: "16ab80a6-c4a1-4df0-af42-6711e90313c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.570588 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.570628 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jz7g\" (UniqueName: \"kubernetes.io/projected/16ab80a6-c4a1-4df0-af42-6711e90313c0-kube-api-access-4jz7g\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.570645 5002 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.570661 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.570673 5002 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab80a6-c4a1-4df0-af42-6711e90313c0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.731074 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8" path="/var/lib/kubelet/pods/1a145b9e-3b5d-4307-a6a6-f8c0ce3afab8/volumes" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.750103 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d10fce14-aa5a-466f-948a-7787f2aa8a76","Type":"ContainerStarted","Data":"a60e71f8e6ec08c495dfb3d108ee41d43ea02cc64d4998420722187d26bdef35"} Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.750154 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"d10fce14-aa5a-466f-948a-7787f2aa8a76","Type":"ContainerStarted","Data":"7324cd9cf5e170dce4589482b8b66a1fe7f38be1dbe740ed497a4df168b13057"} Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.755208 5002 generic.go:334] "Generic (PLEG): container finished" podID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerID="c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437" exitCode=0 Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.755244 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerDied","Data":"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437"} Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.755267 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"16ab80a6-c4a1-4df0-af42-6711e90313c0","Type":"ContainerDied","Data":"91bde49da7eebae6ee7621634c7a6a769811252ccf2a55a43039b4afffd47a79"} Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.755287 5002 scope.go:117] "RemoveContainer" containerID="96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.755288 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.777973 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.777956502 podStartE2EDuration="2.777956502s" podCreationTimestamp="2025-10-14 08:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:44:21.776781961 +0000 UTC m=+3194.758021433" watchObservedRunningTime="2025-10-14 08:44:21.777956502 +0000 UTC m=+3194.759195954" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.805283 5002 scope.go:117] "RemoveContainer" containerID="c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.805468 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.812381 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.844301 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:21 crc kubenswrapper[5002]: E1014 08:44:21.844765 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="manila-scheduler" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.844787 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="manila-scheduler" Oct 14 08:44:21 crc kubenswrapper[5002]: E1014 08:44:21.844817 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="probe" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.844827 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="probe" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.845025 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="manila-scheduler" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.845043 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" containerName="probe" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.854494 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.861697 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.867232 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.877513 5002 scope.go:117] "RemoveContainer" containerID="96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28" Oct 14 08:44:21 crc kubenswrapper[5002]: E1014 08:44:21.878854 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28\": container with ID starting with 96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28 not found: ID does not exist" containerID="96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.878881 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28"} err="failed to get container status \"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28\": rpc error: code = NotFound desc = could not find container \"96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28\": container with ID starting with 96a16dab48b4c687958e7412846e4d4b33166f5d3e4925880eec1a44b23dab28 not found: ID does not exist" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.878899 5002 scope.go:117] "RemoveContainer" containerID="c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437" Oct 14 08:44:21 crc kubenswrapper[5002]: E1014 08:44:21.880235 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437\": container with ID starting with c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437 not found: ID does not exist" containerID="c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.880262 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437"} err="failed to get container status \"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437\": rpc error: code = NotFound desc = could not find container \"c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437\": container with ID starting with c2a804ed036ebd0671b202473330900e73d16cb5ef824992373ec61ab4ba3437 not found: ID does not exist" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.995246 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5tf6\" (UniqueName: \"kubernetes.io/projected/0bdd233c-7fae-4db7-8056-a218f6da8754-kube-api-access-z5tf6\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.995312 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0bdd233c-7fae-4db7-8056-a218f6da8754-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.995373 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.995422 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.995897 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-scripts\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:21 crc kubenswrapper[5002]: I1014 08:44:21.996134 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.097822 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.097915 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.097973 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-scripts\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.098019 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.098115 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5tf6\" (UniqueName: \"kubernetes.io/projected/0bdd233c-7fae-4db7-8056-a218f6da8754-kube-api-access-z5tf6\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.098155 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0bdd233c-7fae-4db7-8056-a218f6da8754-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.098253 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0bdd233c-7fae-4db7-8056-a218f6da8754-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.104114 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.104771 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-config-data\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.106761 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-scripts\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.107978 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bdd233c-7fae-4db7-8056-a218f6da8754-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.115729 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5tf6\" (UniqueName: \"kubernetes.io/projected/0bdd233c-7fae-4db7-8056-a218f6da8754-kube-api-access-z5tf6\") pod \"manila-scheduler-0\" (UID: \"0bdd233c-7fae-4db7-8056-a218f6da8754\") " pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.193766 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.664775 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 08:44:22 crc kubenswrapper[5002]: I1014 08:44:22.771644 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"0bdd233c-7fae-4db7-8056-a218f6da8754","Type":"ContainerStarted","Data":"7883cf423a61fb05159c193c7615cdfa71beafcb0b6f036001fe6f4207114a79"} Oct 14 08:44:23 crc kubenswrapper[5002]: I1014 08:44:23.181140 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 14 08:44:23 crc kubenswrapper[5002]: I1014 08:44:23.741646 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ab80a6-c4a1-4df0-af42-6711e90313c0" path="/var/lib/kubelet/pods/16ab80a6-c4a1-4df0-af42-6711e90313c0/volumes" Oct 14 08:44:23 crc kubenswrapper[5002]: I1014 08:44:23.780307 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"0bdd233c-7fae-4db7-8056-a218f6da8754","Type":"ContainerStarted","Data":"9f15423c4bc8d9f7c571e88737e0a71dbc21e4266a23b38a00c0b14a43289f12"} Oct 14 08:44:23 crc kubenswrapper[5002]: I1014 08:44:23.780371 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"0bdd233c-7fae-4db7-8056-a218f6da8754","Type":"ContainerStarted","Data":"f5750f152ec2d3d00c9627712e73edcfcc02660ecd041774eece288ba8139dda"} Oct 14 08:44:23 crc kubenswrapper[5002]: I1014 08:44:23.801183 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.801165509 podStartE2EDuration="2.801165509s" podCreationTimestamp="2025-10-14 08:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:44:23.799614068 +0000 UTC m=+3196.780853540" watchObservedRunningTime="2025-10-14 08:44:23.801165509 +0000 UTC m=+3196.782404961" Oct 14 08:44:30 crc kubenswrapper[5002]: I1014 08:44:30.138043 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 14 08:44:32 crc kubenswrapper[5002]: I1014 08:44:32.194528 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 14 08:44:41 crc kubenswrapper[5002]: I1014 08:44:41.540192 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 14 08:44:42 crc kubenswrapper[5002]: I1014 08:44:42.338886 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 08:44:43 crc kubenswrapper[5002]: I1014 08:44:43.769064 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.204349 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm"] Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.207642 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.212271 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.212423 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.218980 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm"] Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.335568 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.335641 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.335818 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9zsf\" (UniqueName: \"kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.437911 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9zsf\" (UniqueName: \"kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.438063 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.438210 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.439952 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.448422 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.472962 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9zsf\" (UniqueName: \"kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf\") pod \"collect-profiles-29340525-tr2tm\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:00 crc kubenswrapper[5002]: I1014 08:45:00.549626 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:01 crc kubenswrapper[5002]: I1014 08:45:01.091198 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm"] Oct 14 08:45:01 crc kubenswrapper[5002]: I1014 08:45:01.258540 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" event={"ID":"853cca8b-4f20-40a5-91ae-50e50417395f","Type":"ContainerStarted","Data":"8f24a2d179367023dd4cad8963bc5d0f0a8b10dc19cce23aa389a84a7878b157"} Oct 14 08:45:02 crc kubenswrapper[5002]: I1014 08:45:02.273355 5002 generic.go:334] "Generic (PLEG): container finished" podID="853cca8b-4f20-40a5-91ae-50e50417395f" containerID="6391c8c9491417a713d8da78dbe3e379d8d87aba5a3f1f23b631364ee5953b1b" exitCode=0 Oct 14 08:45:02 crc kubenswrapper[5002]: I1014 08:45:02.273454 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" event={"ID":"853cca8b-4f20-40a5-91ae-50e50417395f","Type":"ContainerDied","Data":"6391c8c9491417a713d8da78dbe3e379d8d87aba5a3f1f23b631364ee5953b1b"} Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.676258 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.808132 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9zsf\" (UniqueName: \"kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf\") pod \"853cca8b-4f20-40a5-91ae-50e50417395f\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.808366 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume\") pod \"853cca8b-4f20-40a5-91ae-50e50417395f\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.808478 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume\") pod \"853cca8b-4f20-40a5-91ae-50e50417395f\" (UID: \"853cca8b-4f20-40a5-91ae-50e50417395f\") " Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.810967 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume" (OuterVolumeSpecName: "config-volume") pod "853cca8b-4f20-40a5-91ae-50e50417395f" (UID: "853cca8b-4f20-40a5-91ae-50e50417395f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.814669 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf" (OuterVolumeSpecName: "kube-api-access-c9zsf") pod "853cca8b-4f20-40a5-91ae-50e50417395f" (UID: "853cca8b-4f20-40a5-91ae-50e50417395f"). InnerVolumeSpecName "kube-api-access-c9zsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.817042 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "853cca8b-4f20-40a5-91ae-50e50417395f" (UID: "853cca8b-4f20-40a5-91ae-50e50417395f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.911172 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/853cca8b-4f20-40a5-91ae-50e50417395f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.911214 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9zsf\" (UniqueName: \"kubernetes.io/projected/853cca8b-4f20-40a5-91ae-50e50417395f-kube-api-access-c9zsf\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[5002]: I1014 08:45:03.911229 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/853cca8b-4f20-40a5-91ae-50e50417395f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:04 crc kubenswrapper[5002]: I1014 08:45:04.305697 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" event={"ID":"853cca8b-4f20-40a5-91ae-50e50417395f","Type":"ContainerDied","Data":"8f24a2d179367023dd4cad8963bc5d0f0a8b10dc19cce23aa389a84a7878b157"} Oct 14 08:45:04 crc kubenswrapper[5002]: I1014 08:45:04.305755 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm" Oct 14 08:45:04 crc kubenswrapper[5002]: I1014 08:45:04.305908 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f24a2d179367023dd4cad8963bc5d0f0a8b10dc19cce23aa389a84a7878b157" Oct 14 08:45:04 crc kubenswrapper[5002]: I1014 08:45:04.805034 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8"] Oct 14 08:45:04 crc kubenswrapper[5002]: I1014 08:45:04.815996 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-cc5x8"] Oct 14 08:45:05 crc kubenswrapper[5002]: I1014 08:45:05.743378 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a0268a0-bec8-43f1-864d-c54e0151fb34" path="/var/lib/kubelet/pods/2a0268a0-bec8-43f1-864d-c54e0151fb34/volumes" Oct 14 08:45:25 crc kubenswrapper[5002]: I1014 08:45:25.627820 5002 scope.go:117] "RemoveContainer" containerID="df850e6f4af91ab4d897f720c99f81649c55f21fbbbedb1ea6455e36220cc3af" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.180342 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd554b564-6djql"] Oct 14 08:45:36 crc kubenswrapper[5002]: E1014 08:45:36.181170 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="853cca8b-4f20-40a5-91ae-50e50417395f" containerName="collect-profiles" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.181182 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="853cca8b-4f20-40a5-91ae-50e50417395f" containerName="collect-profiles" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.181389 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="853cca8b-4f20-40a5-91ae-50e50417395f" containerName="collect-profiles" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.182989 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.209089 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd554b564-6djql"] Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.302015 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psvb7\" (UniqueName: \"kubernetes.io/projected/41ec772c-a692-4e45-90f1-5466203d5587-kube-api-access-psvb7\") pod \"openstack-operator-controller-operator-bd554b564-6djql\" (UID: \"41ec772c-a692-4e45-90f1-5466203d5587\") " pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.403634 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psvb7\" (UniqueName: \"kubernetes.io/projected/41ec772c-a692-4e45-90f1-5466203d5587-kube-api-access-psvb7\") pod \"openstack-operator-controller-operator-bd554b564-6djql\" (UID: \"41ec772c-a692-4e45-90f1-5466203d5587\") " pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.428069 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psvb7\" (UniqueName: \"kubernetes.io/projected/41ec772c-a692-4e45-90f1-5466203d5587-kube-api-access-psvb7\") pod \"openstack-operator-controller-operator-bd554b564-6djql\" (UID: \"41ec772c-a692-4e45-90f1-5466203d5587\") " pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:36 crc kubenswrapper[5002]: I1014 08:45:36.504203 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.039687 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd554b564-6djql"] Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.659489 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" event={"ID":"41ec772c-a692-4e45-90f1-5466203d5587","Type":"ContainerStarted","Data":"0fcc44ce69f5d391c1a356526f5de33233164f454a62bc14411a756f8aa73902"} Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.660148 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" event={"ID":"41ec772c-a692-4e45-90f1-5466203d5587","Type":"ContainerStarted","Data":"415b152a56a0b319d34e332eace4445a50c3261ac4ebc806db088c8d5df927c4"} Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.660162 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" event={"ID":"41ec772c-a692-4e45-90f1-5466203d5587","Type":"ContainerStarted","Data":"1df121afec28089f90b37957eb498f97b3f2402e90d012a749bca865c867aaf2"} Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.660176 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:37 crc kubenswrapper[5002]: I1014 08:45:37.701862 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" podStartSLOduration=1.701824841 podStartE2EDuration="1.701824841s" podCreationTimestamp="2025-10-14 08:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:45:37.692954533 +0000 UTC m=+3270.674193995" watchObservedRunningTime="2025-10-14 08:45:37.701824841 +0000 UTC m=+3270.683064293" Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.510178 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-bd554b564-6djql" Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.642933 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.643330 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="operator" containerID="cri-o://4843b6ff5cad6ca4bd2edfa89e75ba3f55cf28f5ad0b97de33d73fc582ecb619" gracePeriod=10 Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.643403 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="kube-rbac-proxy" containerID="cri-o://9352e0d127ce543c017c1156b3d81e8c54c20be48de0865c50640db9feb31632" gracePeriod=10 Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.790887 5002 generic.go:334] "Generic (PLEG): container finished" podID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerID="9352e0d127ce543c017c1156b3d81e8c54c20be48de0865c50640db9feb31632" exitCode=0 Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.790994 5002 generic.go:334] "Generic (PLEG): container finished" podID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerID="4843b6ff5cad6ca4bd2edfa89e75ba3f55cf28f5ad0b97de33d73fc582ecb619" exitCode=0 Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.791038 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerDied","Data":"9352e0d127ce543c017c1156b3d81e8c54c20be48de0865c50640db9feb31632"} Oct 14 08:45:46 crc kubenswrapper[5002]: I1014 08:45:46.791108 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerDied","Data":"4843b6ff5cad6ca4bd2edfa89e75ba3f55cf28f5ad0b97de33d73fc582ecb619"} Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.140562 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.302963 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9stnv\" (UniqueName: \"kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv\") pod \"78216670-be87-44b5-be7e-b3e98f8e85e7\" (UID: \"78216670-be87-44b5-be7e-b3e98f8e85e7\") " Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.309310 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv" (OuterVolumeSpecName: "kube-api-access-9stnv") pod "78216670-be87-44b5-be7e-b3e98f8e85e7" (UID: "78216670-be87-44b5-be7e-b3e98f8e85e7"). InnerVolumeSpecName "kube-api-access-9stnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.405661 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9stnv\" (UniqueName: \"kubernetes.io/projected/78216670-be87-44b5-be7e-b3e98f8e85e7-kube-api-access-9stnv\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.804930 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" event={"ID":"78216670-be87-44b5-be7e-b3e98f8e85e7","Type":"ContainerDied","Data":"7bbe82ee5a8b29d536ed728ee1ba55a62eb6d5027cde5ab6565af2260fcc7dee"} Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.805007 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.805031 5002 scope.go:117] "RemoveContainer" containerID="9352e0d127ce543c017c1156b3d81e8c54c20be48de0865c50640db9feb31632" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.839812 5002 scope.go:117] "RemoveContainer" containerID="4843b6ff5cad6ca4bd2edfa89e75ba3f55cf28f5ad0b97de33d73fc582ecb619" Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.854070 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:45:47 crc kubenswrapper[5002]: I1014 08:45:47.862997 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-k6rlc"] Oct 14 08:45:49 crc kubenswrapper[5002]: I1014 08:45:49.741980 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" path="/var/lib/kubelet/pods/78216670-be87-44b5-be7e-b3e98f8e85e7/volumes" Oct 14 08:46:09 crc kubenswrapper[5002]: I1014 08:46:09.218260 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:46:09 crc kubenswrapper[5002]: I1014 08:46:09.218688 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.074170 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v"] Oct 14 08:46:20 crc kubenswrapper[5002]: E1014 08:46:20.075089 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="kube-rbac-proxy" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.075103 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="kube-rbac-proxy" Oct 14 08:46:20 crc kubenswrapper[5002]: E1014 08:46:20.075126 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="operator" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.075133 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="operator" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.075311 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="kube-rbac-proxy" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.075338 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="78216670-be87-44b5-be7e-b3e98f8e85e7" containerName="operator" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.076262 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.087473 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v"] Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.219285 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68c57\" (UniqueName: \"kubernetes.io/projected/cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f-kube-api-access-68c57\") pod \"test-operator-controller-manager-c475b9f8d-pbb6v\" (UID: \"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f\") " pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.321175 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68c57\" (UniqueName: \"kubernetes.io/projected/cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f-kube-api-access-68c57\") pod \"test-operator-controller-manager-c475b9f8d-pbb6v\" (UID: \"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f\") " pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.345735 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68c57\" (UniqueName: \"kubernetes.io/projected/cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f-kube-api-access-68c57\") pod \"test-operator-controller-manager-c475b9f8d-pbb6v\" (UID: \"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f\") " pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.399715 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:20 crc kubenswrapper[5002]: I1014 08:46:20.890551 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v"] Oct 14 08:46:21 crc kubenswrapper[5002]: I1014 08:46:21.174216 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" event={"ID":"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f","Type":"ContainerStarted","Data":"44a83af2a2886a95b6ca8d08de13344228f539f64f09fce1d3d468c5391f8e85"} Oct 14 08:46:22 crc kubenswrapper[5002]: I1014 08:46:22.209161 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" event={"ID":"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f","Type":"ContainerStarted","Data":"6b304cb83f1fc0f00fd9b1c3b304e83f4d2c1ecb812c0ce3158d25cb7bfd99de"} Oct 14 08:46:22 crc kubenswrapper[5002]: I1014 08:46:22.209437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" event={"ID":"cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f","Type":"ContainerStarted","Data":"b12e575639359dfde082c65f683dd851ae82aa62d094f0dfec498f6173e24aca"} Oct 14 08:46:22 crc kubenswrapper[5002]: I1014 08:46:22.209549 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:22 crc kubenswrapper[5002]: I1014 08:46:22.252048 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" podStartSLOduration=1.411921642 podStartE2EDuration="2.252021884s" podCreationTimestamp="2025-10-14 08:46:20 +0000 UTC" firstStartedPulling="2025-10-14 08:46:20.903181639 +0000 UTC m=+3313.884421121" lastFinishedPulling="2025-10-14 08:46:21.743281911 +0000 UTC m=+3314.724521363" observedRunningTime="2025-10-14 08:46:22.226986203 +0000 UTC m=+3315.208225675" watchObservedRunningTime="2025-10-14 08:46:22.252021884 +0000 UTC m=+3315.233261336" Oct 14 08:46:30 crc kubenswrapper[5002]: I1014 08:46:30.404004 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-c475b9f8d-pbb6v" Oct 14 08:46:30 crc kubenswrapper[5002]: I1014 08:46:30.497761 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:46:30 crc kubenswrapper[5002]: I1014 08:46:30.498010 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="kube-rbac-proxy" containerID="cri-o://6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" gracePeriod=10 Oct 14 08:46:30 crc kubenswrapper[5002]: I1014 08:46:30.498139 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="manager" containerID="cri-o://e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" gracePeriod=10 Oct 14 08:46:30 crc kubenswrapper[5002]: I1014 08:46:30.938005 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.094351 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98fqt\" (UniqueName: \"kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt\") pod \"e4e3eea0-46ac-4106-bf62-6cd798e2f96b\" (UID: \"e4e3eea0-46ac-4106-bf62-6cd798e2f96b\") " Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.100655 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt" (OuterVolumeSpecName: "kube-api-access-98fqt") pod "e4e3eea0-46ac-4106-bf62-6cd798e2f96b" (UID: "e4e3eea0-46ac-4106-bf62-6cd798e2f96b"). InnerVolumeSpecName "kube-api-access-98fqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.196964 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98fqt\" (UniqueName: \"kubernetes.io/projected/e4e3eea0-46ac-4106-bf62-6cd798e2f96b-kube-api-access-98fqt\") on node \"crc\" DevicePath \"\"" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303334 5002 generic.go:334] "Generic (PLEG): container finished" podID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerID="e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" exitCode=0 Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303390 5002 generic.go:334] "Generic (PLEG): container finished" podID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerID="6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" exitCode=0 Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303394 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303431 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerDied","Data":"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da"} Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303502 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerDied","Data":"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b"} Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303528 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn" event={"ID":"e4e3eea0-46ac-4106-bf62-6cd798e2f96b","Type":"ContainerDied","Data":"c2bc9942ad0793643bee9238e41da6673444abcb51aee18d6e32561967e7ae9c"} Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.303561 5002 scope.go:117] "RemoveContainer" containerID="e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.328157 5002 scope.go:117] "RemoveContainer" containerID="6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.338363 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.346834 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-xd9hn"] Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.351668 5002 scope.go:117] "RemoveContainer" containerID="e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" Oct 14 08:46:31 crc kubenswrapper[5002]: E1014 08:46:31.352131 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da\": container with ID starting with e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da not found: ID does not exist" containerID="e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.352169 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da"} err="failed to get container status \"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da\": rpc error: code = NotFound desc = could not find container \"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da\": container with ID starting with e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da not found: ID does not exist" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.352189 5002 scope.go:117] "RemoveContainer" containerID="6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" Oct 14 08:46:31 crc kubenswrapper[5002]: E1014 08:46:31.352892 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b\": container with ID starting with 6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b not found: ID does not exist" containerID="6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.352917 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b"} err="failed to get container status \"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b\": rpc error: code = NotFound desc = could not find container \"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b\": container with ID starting with 6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b not found: ID does not exist" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.352931 5002 scope.go:117] "RemoveContainer" containerID="e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.353178 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da"} err="failed to get container status \"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da\": rpc error: code = NotFound desc = could not find container \"e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da\": container with ID starting with e6d604b8851bcdff01a27d9b46d82c901906a58b35bafb395521d19c119c01da not found: ID does not exist" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.353196 5002 scope.go:117] "RemoveContainer" containerID="6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.353427 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b"} err="failed to get container status \"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b\": rpc error: code = NotFound desc = could not find container \"6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b\": container with ID starting with 6d2c531006055a4677e374f6568d49daac869a05f089d4f186e8bca1fd03f91b not found: ID does not exist" Oct 14 08:46:31 crc kubenswrapper[5002]: I1014 08:46:31.732720 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" path="/var/lib/kubelet/pods/e4e3eea0-46ac-4106-bf62-6cd798e2f96b/volumes" Oct 14 08:46:39 crc kubenswrapper[5002]: I1014 08:46:39.220250 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:46:39 crc kubenswrapper[5002]: I1014 08:46:39.221899 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.554020 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:46:59 crc kubenswrapper[5002]: E1014 08:46:59.555524 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="manager" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.555558 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="manager" Oct 14 08:46:59 crc kubenswrapper[5002]: E1014 08:46:59.555601 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="kube-rbac-proxy" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.555619 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="kube-rbac-proxy" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.556120 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="kube-rbac-proxy" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.556224 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4e3eea0-46ac-4106-bf62-6cd798e2f96b" containerName="manager" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.559460 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.570892 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.635288 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.635463 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s4f6\" (UniqueName: \"kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.635617 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.737113 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.737265 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s4f6\" (UniqueName: \"kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.737438 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.737710 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.738254 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.764382 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s4f6\" (UniqueName: \"kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6\") pod \"redhat-operators-44gtc\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:46:59 crc kubenswrapper[5002]: I1014 08:46:59.884313 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:00 crc kubenswrapper[5002]: I1014 08:47:00.333174 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:47:00 crc kubenswrapper[5002]: I1014 08:47:00.652224 5002 generic.go:334] "Generic (PLEG): container finished" podID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerID="728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53" exitCode=0 Oct 14 08:47:00 crc kubenswrapper[5002]: I1014 08:47:00.652371 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerDied","Data":"728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53"} Oct 14 08:47:00 crc kubenswrapper[5002]: I1014 08:47:00.653418 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerStarted","Data":"0870507d7ff44415dc009901b155389754e3ffada736da53308b4f62452d0799"} Oct 14 08:47:02 crc kubenswrapper[5002]: I1014 08:47:02.676776 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerStarted","Data":"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99"} Oct 14 08:47:04 crc kubenswrapper[5002]: I1014 08:47:04.703437 5002 generic.go:334] "Generic (PLEG): container finished" podID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerID="4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99" exitCode=0 Oct 14 08:47:04 crc kubenswrapper[5002]: I1014 08:47:04.703519 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerDied","Data":"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99"} Oct 14 08:47:05 crc kubenswrapper[5002]: I1014 08:47:05.716161 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerStarted","Data":"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63"} Oct 14 08:47:05 crc kubenswrapper[5002]: I1014 08:47:05.744379 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-44gtc" podStartSLOduration=2.100063043 podStartE2EDuration="6.744340629s" podCreationTimestamp="2025-10-14 08:46:59 +0000 UTC" firstStartedPulling="2025-10-14 08:47:00.653989039 +0000 UTC m=+3353.635228491" lastFinishedPulling="2025-10-14 08:47:05.298266625 +0000 UTC m=+3358.279506077" observedRunningTime="2025-10-14 08:47:05.739635483 +0000 UTC m=+3358.720874955" watchObservedRunningTime="2025-10-14 08:47:05.744340629 +0000 UTC m=+3358.725580111" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.218109 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.218701 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.218760 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.219718 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.219802 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a" gracePeriod=600 Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.768077 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a" exitCode=0 Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.768144 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a"} Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.768710 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207"} Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.768747 5002 scope.go:117] "RemoveContainer" containerID="b1d20f63944506329b8e3159a62cec13b38cb0c5d870d2096fe71f8a0626a3ca" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.884734 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:09 crc kubenswrapper[5002]: I1014 08:47:09.884782 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:10 crc kubenswrapper[5002]: I1014 08:47:10.950861 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-44gtc" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="registry-server" probeResult="failure" output=< Oct 14 08:47:10 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 08:47:10 crc kubenswrapper[5002]: > Oct 14 08:47:19 crc kubenswrapper[5002]: I1014 08:47:19.935443 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:20 crc kubenswrapper[5002]: I1014 08:47:20.012456 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:20 crc kubenswrapper[5002]: I1014 08:47:20.182814 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:47:21 crc kubenswrapper[5002]: I1014 08:47:21.917522 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-44gtc" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="registry-server" containerID="cri-o://94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63" gracePeriod=2 Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.357499 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.484640 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8s4f6\" (UniqueName: \"kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6\") pod \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.484737 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content\") pod \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.484792 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities\") pod \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\" (UID: \"c727cccc-a16a-4d72-84d5-2dfc4ae968cb\") " Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.485825 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities" (OuterVolumeSpecName: "utilities") pod "c727cccc-a16a-4d72-84d5-2dfc4ae968cb" (UID: "c727cccc-a16a-4d72-84d5-2dfc4ae968cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.500294 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6" (OuterVolumeSpecName: "kube-api-access-8s4f6") pod "c727cccc-a16a-4d72-84d5-2dfc4ae968cb" (UID: "c727cccc-a16a-4d72-84d5-2dfc4ae968cb"). InnerVolumeSpecName "kube-api-access-8s4f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.570554 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c727cccc-a16a-4d72-84d5-2dfc4ae968cb" (UID: "c727cccc-a16a-4d72-84d5-2dfc4ae968cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.587025 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8s4f6\" (UniqueName: \"kubernetes.io/projected/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-kube-api-access-8s4f6\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.587047 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.587060 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c727cccc-a16a-4d72-84d5-2dfc4ae968cb-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.931992 5002 generic.go:334] "Generic (PLEG): container finished" podID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerID="94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63" exitCode=0 Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.932065 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-44gtc" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.932069 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerDied","Data":"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63"} Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.932121 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-44gtc" event={"ID":"c727cccc-a16a-4d72-84d5-2dfc4ae968cb","Type":"ContainerDied","Data":"0870507d7ff44415dc009901b155389754e3ffada736da53308b4f62452d0799"} Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.932174 5002 scope.go:117] "RemoveContainer" containerID="94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63" Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.972854 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.981960 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-44gtc"] Oct 14 08:47:22 crc kubenswrapper[5002]: I1014 08:47:22.983354 5002 scope.go:117] "RemoveContainer" containerID="4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.022172 5002 scope.go:117] "RemoveContainer" containerID="728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.057257 5002 scope.go:117] "RemoveContainer" containerID="94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63" Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.057670 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63\": container with ID starting with 94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63 not found: ID does not exist" containerID="94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.057698 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63"} err="failed to get container status \"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63\": rpc error: code = NotFound desc = could not find container \"94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63\": container with ID starting with 94926b7ca63034f7e5db7cdab069a48d2dfa96efdd8069204fa9accf32915d63 not found: ID does not exist" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.057720 5002 scope.go:117] "RemoveContainer" containerID="4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99" Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.058080 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99\": container with ID starting with 4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99 not found: ID does not exist" containerID="4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.058102 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99"} err="failed to get container status \"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99\": rpc error: code = NotFound desc = could not find container \"4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99\": container with ID starting with 4bec853868f653e31f312991b0342305f2a0626ebc4150d5a1da432c45e0cd99 not found: ID does not exist" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.058114 5002 scope.go:117] "RemoveContainer" containerID="728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53" Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.058530 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53\": container with ID starting with 728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53 not found: ID does not exist" containerID="728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.058560 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53"} err="failed to get container status \"728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53\": rpc error: code = NotFound desc = could not find container \"728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53\": container with ID starting with 728f1291ef6ff04659e45148bd70ab74235a3631c32a6780763b457c847d5d53 not found: ID does not exist" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.184728 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.185211 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="registry-server" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.185234 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="registry-server" Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.185250 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="extract-utilities" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.185256 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="extract-utilities" Oct 14 08:47:23 crc kubenswrapper[5002]: E1014 08:47:23.185276 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="extract-content" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.185282 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="extract-content" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.185450 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" containerName="registry-server" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.186925 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.212590 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.307064 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94xkj\" (UniqueName: \"kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.307589 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.307964 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.409932 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94xkj\" (UniqueName: \"kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.410024 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.410108 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.410633 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.410665 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.433778 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94xkj\" (UniqueName: \"kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj\") pod \"certified-operators-mtshq\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.505131 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.735688 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c727cccc-a16a-4d72-84d5-2dfc4ae968cb" path="/var/lib/kubelet/pods/c727cccc-a16a-4d72-84d5-2dfc4ae968cb/volumes" Oct 14 08:47:23 crc kubenswrapper[5002]: I1014 08:47:23.991309 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.220944 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.223300 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.227581 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.332311 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l764d\" (UniqueName: \"kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.332424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.332481 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.433916 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l764d\" (UniqueName: \"kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.434014 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.434063 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.434575 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.434605 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.471821 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l764d\" (UniqueName: \"kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d\") pod \"community-operators-blc2s\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.555394 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.952585 5002 generic.go:334] "Generic (PLEG): container finished" podID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerID="6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e" exitCode=0 Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.952676 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerDied","Data":"6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e"} Oct 14 08:47:24 crc kubenswrapper[5002]: I1014 08:47:24.952894 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerStarted","Data":"76c58026eef761fae37f1c9efa08a16f24451304c1b181d2bc12ac54160a8169"} Oct 14 08:47:25 crc kubenswrapper[5002]: I1014 08:47:25.039427 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:25 crc kubenswrapper[5002]: W1014 08:47:25.045412 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54147e84_e855_4b14_ba8d_704fa2243a77.slice/crio-5629af0c24c44ea58fd5b2a0fd43f14c6c44e615be804a3c9de19fbd327a0320 WatchSource:0}: Error finding container 5629af0c24c44ea58fd5b2a0fd43f14c6c44e615be804a3c9de19fbd327a0320: Status 404 returned error can't find the container with id 5629af0c24c44ea58fd5b2a0fd43f14c6c44e615be804a3c9de19fbd327a0320 Oct 14 08:47:25 crc kubenswrapper[5002]: I1014 08:47:25.963492 5002 generic.go:334] "Generic (PLEG): container finished" podID="54147e84-e855-4b14-ba8d-704fa2243a77" containerID="d460c45a183461650e9d4ce1638ad9641cd151c33d476d46bdf99e3d8f894478" exitCode=0 Oct 14 08:47:25 crc kubenswrapper[5002]: I1014 08:47:25.963573 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerDied","Data":"d460c45a183461650e9d4ce1638ad9641cd151c33d476d46bdf99e3d8f894478"} Oct 14 08:47:25 crc kubenswrapper[5002]: I1014 08:47:25.963822 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerStarted","Data":"5629af0c24c44ea58fd5b2a0fd43f14c6c44e615be804a3c9de19fbd327a0320"} Oct 14 08:47:26 crc kubenswrapper[5002]: I1014 08:47:26.986021 5002 generic.go:334] "Generic (PLEG): container finished" podID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerID="5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b" exitCode=0 Oct 14 08:47:26 crc kubenswrapper[5002]: I1014 08:47:26.986093 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerDied","Data":"5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b"} Oct 14 08:47:28 crc kubenswrapper[5002]: I1014 08:47:28.010166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerStarted","Data":"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127"} Oct 14 08:47:28 crc kubenswrapper[5002]: I1014 08:47:28.012026 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerStarted","Data":"f4fe8d48c61ff8f3631834177a7c6df1f6e040f8ec4acae57e68f01ecbf5784d"} Oct 14 08:47:28 crc kubenswrapper[5002]: I1014 08:47:28.027699 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mtshq" podStartSLOduration=2.229562742 podStartE2EDuration="5.027678015s" podCreationTimestamp="2025-10-14 08:47:23 +0000 UTC" firstStartedPulling="2025-10-14 08:47:24.955091587 +0000 UTC m=+3377.936331049" lastFinishedPulling="2025-10-14 08:47:27.75320686 +0000 UTC m=+3380.734446322" observedRunningTime="2025-10-14 08:47:28.026570025 +0000 UTC m=+3381.007809497" watchObservedRunningTime="2025-10-14 08:47:28.027678015 +0000 UTC m=+3381.008917497" Oct 14 08:47:29 crc kubenswrapper[5002]: I1014 08:47:29.022041 5002 generic.go:334] "Generic (PLEG): container finished" podID="54147e84-e855-4b14-ba8d-704fa2243a77" containerID="f4fe8d48c61ff8f3631834177a7c6df1f6e040f8ec4acae57e68f01ecbf5784d" exitCode=0 Oct 14 08:47:29 crc kubenswrapper[5002]: I1014 08:47:29.022102 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerDied","Data":"f4fe8d48c61ff8f3631834177a7c6df1f6e040f8ec4acae57e68f01ecbf5784d"} Oct 14 08:47:30 crc kubenswrapper[5002]: I1014 08:47:30.034760 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerStarted","Data":"b5ab0f7c5db4604fea9556220d5343ec520c87b08a8380d2a3cc59e3c7fbeddd"} Oct 14 08:47:33 crc kubenswrapper[5002]: I1014 08:47:33.505966 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:33 crc kubenswrapper[5002]: I1014 08:47:33.506457 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:33 crc kubenswrapper[5002]: I1014 08:47:33.564994 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:33 crc kubenswrapper[5002]: I1014 08:47:33.597155 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-blc2s" podStartSLOduration=6.081351769 podStartE2EDuration="9.597132843s" podCreationTimestamp="2025-10-14 08:47:24 +0000 UTC" firstStartedPulling="2025-10-14 08:47:25.9653467 +0000 UTC m=+3378.946586162" lastFinishedPulling="2025-10-14 08:47:29.481127774 +0000 UTC m=+3382.462367236" observedRunningTime="2025-10-14 08:47:30.061171807 +0000 UTC m=+3383.042411309" watchObservedRunningTime="2025-10-14 08:47:33.597132843 +0000 UTC m=+3386.578372305" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.118011 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.188344 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.191025 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.199314 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.341364 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.342028 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mck\" (UniqueName: \"kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.342393 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.443885 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.444070 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mck\" (UniqueName: \"kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.444190 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.444444 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.444547 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.471513 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mck\" (UniqueName: \"kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck\") pod \"redhat-marketplace-58hpb\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.526719 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.556527 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.558446 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:34 crc kubenswrapper[5002]: I1014 08:47:34.611475 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:35 crc kubenswrapper[5002]: I1014 08:47:35.021763 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:35 crc kubenswrapper[5002]: W1014 08:47:35.025685 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e78d8fe_ae4f_4246_abf8_a6a21441fe37.slice/crio-86aba30ae1501106806bba2a1ccfecf20799a676486ea722721fbb369285b2c3 WatchSource:0}: Error finding container 86aba30ae1501106806bba2a1ccfecf20799a676486ea722721fbb369285b2c3: Status 404 returned error can't find the container with id 86aba30ae1501106806bba2a1ccfecf20799a676486ea722721fbb369285b2c3 Oct 14 08:47:35 crc kubenswrapper[5002]: I1014 08:47:35.097733 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerStarted","Data":"86aba30ae1501106806bba2a1ccfecf20799a676486ea722721fbb369285b2c3"} Oct 14 08:47:35 crc kubenswrapper[5002]: I1014 08:47:35.160085 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:35 crc kubenswrapper[5002]: I1014 08:47:35.979934 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.113153 5002 generic.go:334] "Generic (PLEG): container finished" podID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerID="ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586" exitCode=0 Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.113231 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerDied","Data":"ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586"} Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.113612 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mtshq" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="registry-server" containerID="cri-o://bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127" gracePeriod=2 Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.627364 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.793946 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content\") pod \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.794091 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94xkj\" (UniqueName: \"kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj\") pod \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.794197 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities\") pod \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\" (UID: \"7358f21b-9244-4ecc-ae86-b9d44fb65d4c\") " Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.795957 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities" (OuterVolumeSpecName: "utilities") pod "7358f21b-9244-4ecc-ae86-b9d44fb65d4c" (UID: "7358f21b-9244-4ecc-ae86-b9d44fb65d4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.801246 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj" (OuterVolumeSpecName: "kube-api-access-94xkj") pod "7358f21b-9244-4ecc-ae86-b9d44fb65d4c" (UID: "7358f21b-9244-4ecc-ae86-b9d44fb65d4c"). InnerVolumeSpecName "kube-api-access-94xkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.851703 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7358f21b-9244-4ecc-ae86-b9d44fb65d4c" (UID: "7358f21b-9244-4ecc-ae86-b9d44fb65d4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.897604 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94xkj\" (UniqueName: \"kubernetes.io/projected/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-kube-api-access-94xkj\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.897648 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:36 crc kubenswrapper[5002]: I1014 08:47:36.897666 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7358f21b-9244-4ecc-ae86-b9d44fb65d4c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.126894 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerStarted","Data":"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a"} Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.132905 5002 generic.go:334] "Generic (PLEG): container finished" podID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerID="bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127" exitCode=0 Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.132986 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mtshq" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.132985 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerDied","Data":"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127"} Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.133372 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mtshq" event={"ID":"7358f21b-9244-4ecc-ae86-b9d44fb65d4c","Type":"ContainerDied","Data":"76c58026eef761fae37f1c9efa08a16f24451304c1b181d2bc12ac54160a8169"} Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.133415 5002 scope.go:117] "RemoveContainer" containerID="bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.170073 5002 scope.go:117] "RemoveContainer" containerID="5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.246467 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.256602 5002 scope.go:117] "RemoveContainer" containerID="6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.258254 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mtshq"] Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.295903 5002 scope.go:117] "RemoveContainer" containerID="bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127" Oct 14 08:47:37 crc kubenswrapper[5002]: E1014 08:47:37.296395 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127\": container with ID starting with bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127 not found: ID does not exist" containerID="bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.296430 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127"} err="failed to get container status \"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127\": rpc error: code = NotFound desc = could not find container \"bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127\": container with ID starting with bc2f79324193cb3623bbaa8253ba3a973a64b3fa828dd3b58e1e8a3d18143127 not found: ID does not exist" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.296456 5002 scope.go:117] "RemoveContainer" containerID="5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b" Oct 14 08:47:37 crc kubenswrapper[5002]: E1014 08:47:37.296971 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b\": container with ID starting with 5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b not found: ID does not exist" containerID="5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.296998 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b"} err="failed to get container status \"5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b\": rpc error: code = NotFound desc = could not find container \"5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b\": container with ID starting with 5590169a785d15892baefa35d0172c1bf752bbe63c46f0a8b3728d82f66d3a8b not found: ID does not exist" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.297017 5002 scope.go:117] "RemoveContainer" containerID="6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e" Oct 14 08:47:37 crc kubenswrapper[5002]: E1014 08:47:37.297236 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e\": container with ID starting with 6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e not found: ID does not exist" containerID="6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.297257 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e"} err="failed to get container status \"6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e\": rpc error: code = NotFound desc = could not find container \"6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e\": container with ID starting with 6bc3e4012f34d79a0c179a8a474fc43849ce2226c8f2d188e86248dbeff9a26e not found: ID does not exist" Oct 14 08:47:37 crc kubenswrapper[5002]: I1014 08:47:37.736286 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" path="/var/lib/kubelet/pods/7358f21b-9244-4ecc-ae86-b9d44fb65d4c/volumes" Oct 14 08:47:38 crc kubenswrapper[5002]: I1014 08:47:38.143256 5002 generic.go:334] "Generic (PLEG): container finished" podID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerID="64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a" exitCode=0 Oct 14 08:47:38 crc kubenswrapper[5002]: I1014 08:47:38.143415 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerDied","Data":"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a"} Oct 14 08:47:38 crc kubenswrapper[5002]: I1014 08:47:38.371727 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:38 crc kubenswrapper[5002]: I1014 08:47:38.371998 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-blc2s" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="registry-server" containerID="cri-o://b5ab0f7c5db4604fea9556220d5343ec520c87b08a8380d2a3cc59e3c7fbeddd" gracePeriod=2 Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.164009 5002 generic.go:334] "Generic (PLEG): container finished" podID="54147e84-e855-4b14-ba8d-704fa2243a77" containerID="b5ab0f7c5db4604fea9556220d5343ec520c87b08a8380d2a3cc59e3c7fbeddd" exitCode=0 Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.164104 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerDied","Data":"b5ab0f7c5db4604fea9556220d5343ec520c87b08a8380d2a3cc59e3c7fbeddd"} Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.477914 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.564793 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content\") pod \"54147e84-e855-4b14-ba8d-704fa2243a77\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.564929 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities\") pod \"54147e84-e855-4b14-ba8d-704fa2243a77\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.565163 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l764d\" (UniqueName: \"kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d\") pod \"54147e84-e855-4b14-ba8d-704fa2243a77\" (UID: \"54147e84-e855-4b14-ba8d-704fa2243a77\") " Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.567623 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities" (OuterVolumeSpecName: "utilities") pod "54147e84-e855-4b14-ba8d-704fa2243a77" (UID: "54147e84-e855-4b14-ba8d-704fa2243a77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.572636 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d" (OuterVolumeSpecName: "kube-api-access-l764d") pod "54147e84-e855-4b14-ba8d-704fa2243a77" (UID: "54147e84-e855-4b14-ba8d-704fa2243a77"). InnerVolumeSpecName "kube-api-access-l764d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.630695 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54147e84-e855-4b14-ba8d-704fa2243a77" (UID: "54147e84-e855-4b14-ba8d-704fa2243a77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.668600 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l764d\" (UniqueName: \"kubernetes.io/projected/54147e84-e855-4b14-ba8d-704fa2243a77-kube-api-access-l764d\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.668651 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:39 crc kubenswrapper[5002]: I1014 08:47:39.668671 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54147e84-e855-4b14-ba8d-704fa2243a77-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.176693 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerStarted","Data":"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6"} Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.180440 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-blc2s" event={"ID":"54147e84-e855-4b14-ba8d-704fa2243a77","Type":"ContainerDied","Data":"5629af0c24c44ea58fd5b2a0fd43f14c6c44e615be804a3c9de19fbd327a0320"} Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.180483 5002 scope.go:117] "RemoveContainer" containerID="b5ab0f7c5db4604fea9556220d5343ec520c87b08a8380d2a3cc59e3c7fbeddd" Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.180597 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-blc2s" Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.205580 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-58hpb" podStartSLOduration=3.053562746 podStartE2EDuration="6.205555462s" podCreationTimestamp="2025-10-14 08:47:34 +0000 UTC" firstStartedPulling="2025-10-14 08:47:36.116905256 +0000 UTC m=+3389.098144718" lastFinishedPulling="2025-10-14 08:47:39.268897982 +0000 UTC m=+3392.250137434" observedRunningTime="2025-10-14 08:47:40.194553717 +0000 UTC m=+3393.175793189" watchObservedRunningTime="2025-10-14 08:47:40.205555462 +0000 UTC m=+3393.186794914" Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.214295 5002 scope.go:117] "RemoveContainer" containerID="f4fe8d48c61ff8f3631834177a7c6df1f6e040f8ec4acae57e68f01ecbf5784d" Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.225810 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.235415 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-blc2s"] Oct 14 08:47:40 crc kubenswrapper[5002]: I1014 08:47:40.238070 5002 scope.go:117] "RemoveContainer" containerID="d460c45a183461650e9d4ce1638ad9641cd151c33d476d46bdf99e3d8f894478" Oct 14 08:47:41 crc kubenswrapper[5002]: I1014 08:47:41.741797 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" path="/var/lib/kubelet/pods/54147e84-e855-4b14-ba8d-704fa2243a77/volumes" Oct 14 08:47:44 crc kubenswrapper[5002]: I1014 08:47:44.527620 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:44 crc kubenswrapper[5002]: I1014 08:47:44.528186 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:44 crc kubenswrapper[5002]: I1014 08:47:44.625044 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:45 crc kubenswrapper[5002]: I1014 08:47:45.302343 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:45 crc kubenswrapper[5002]: I1014 08:47:45.574950 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.293273 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-58hpb" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="registry-server" containerID="cri-o://c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6" gracePeriod=2 Oct 14 08:47:47 crc kubenswrapper[5002]: E1014 08:47:47.476876 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e78d8fe_ae4f_4246_abf8_a6a21441fe37.slice/crio-c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e78d8fe_ae4f_4246_abf8_a6a21441fe37.slice/crio-conmon-c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.870446 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.964378 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mck\" (UniqueName: \"kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck\") pod \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.964654 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities\") pod \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.964765 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content\") pod \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\" (UID: \"9e78d8fe-ae4f-4246-abf8-a6a21441fe37\") " Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.966388 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities" (OuterVolumeSpecName: "utilities") pod "9e78d8fe-ae4f-4246-abf8-a6a21441fe37" (UID: "9e78d8fe-ae4f-4246-abf8-a6a21441fe37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.973382 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck" (OuterVolumeSpecName: "kube-api-access-v4mck") pod "9e78d8fe-ae4f-4246-abf8-a6a21441fe37" (UID: "9e78d8fe-ae4f-4246-abf8-a6a21441fe37"). InnerVolumeSpecName "kube-api-access-v4mck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:47:47 crc kubenswrapper[5002]: I1014 08:47:47.978450 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e78d8fe-ae4f-4246-abf8-a6a21441fe37" (UID: "9e78d8fe-ae4f-4246-abf8-a6a21441fe37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.067509 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.067571 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.067602 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mck\" (UniqueName: \"kubernetes.io/projected/9e78d8fe-ae4f-4246-abf8-a6a21441fe37-kube-api-access-v4mck\") on node \"crc\" DevicePath \"\"" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.324084 5002 generic.go:334] "Generic (PLEG): container finished" podID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerID="c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6" exitCode=0 Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.324292 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerDied","Data":"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6"} Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.324384 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58hpb" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.324684 5002 scope.go:117] "RemoveContainer" containerID="c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.324661 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58hpb" event={"ID":"9e78d8fe-ae4f-4246-abf8-a6a21441fe37","Type":"ContainerDied","Data":"86aba30ae1501106806bba2a1ccfecf20799a676486ea722721fbb369285b2c3"} Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.365972 5002 scope.go:117] "RemoveContainer" containerID="64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.388215 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.398355 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-58hpb"] Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.411547 5002 scope.go:117] "RemoveContainer" containerID="ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.439260 5002 scope.go:117] "RemoveContainer" containerID="c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6" Oct 14 08:47:48 crc kubenswrapper[5002]: E1014 08:47:48.439864 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6\": container with ID starting with c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6 not found: ID does not exist" containerID="c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.439922 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6"} err="failed to get container status \"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6\": rpc error: code = NotFound desc = could not find container \"c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6\": container with ID starting with c75de77f277ea9b4bcfb283d68a6e4445d6de2e109d4d231839043c0bfcfdef6 not found: ID does not exist" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.439956 5002 scope.go:117] "RemoveContainer" containerID="64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a" Oct 14 08:47:48 crc kubenswrapper[5002]: E1014 08:47:48.440474 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a\": container with ID starting with 64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a not found: ID does not exist" containerID="64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.440547 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a"} err="failed to get container status \"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a\": rpc error: code = NotFound desc = could not find container \"64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a\": container with ID starting with 64a049a7597afea1c148299be857318605f9a48d66124e7d5e208dc38bf5460a not found: ID does not exist" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.440608 5002 scope.go:117] "RemoveContainer" containerID="ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586" Oct 14 08:47:48 crc kubenswrapper[5002]: E1014 08:47:48.441047 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586\": container with ID starting with ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586 not found: ID does not exist" containerID="ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586" Oct 14 08:47:48 crc kubenswrapper[5002]: I1014 08:47:48.441098 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586"} err="failed to get container status \"ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586\": rpc error: code = NotFound desc = could not find container \"ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586\": container with ID starting with ca9a16a0eb2fe56117dd09266f784e63a403a58dca9833abccd64aa908868586 not found: ID does not exist" Oct 14 08:47:49 crc kubenswrapper[5002]: I1014 08:47:49.739488 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" path="/var/lib/kubelet/pods/9e78d8fe-ae4f-4246-abf8-a6a21441fe37/volumes" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.081537 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084330 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084385 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084444 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084464 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084498 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084516 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084543 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084560 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084590 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084603 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="extract-utilities" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084631 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084645 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084684 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084697 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084725 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084737 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="extract-content" Oct 14 08:48:18 crc kubenswrapper[5002]: E1014 08:48:18.084773 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.084786 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.085243 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e78d8fe-ae4f-4246-abf8-a6a21441fe37" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.085278 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7358f21b-9244-4ecc-ae86-b9d44fb65d4c" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.085339 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="54147e84-e855-4b14-ba8d-704fa2243a77" containerName="registry-server" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.088090 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.100883 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.102732 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.105016 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.105282 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.105403 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7gjvf" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204307 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204371 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204419 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204485 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204590 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204780 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204935 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.204972 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.205059 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.205111 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwjb8\" (UniqueName: \"kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306374 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306429 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306489 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306545 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwjb8\" (UniqueName: \"kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306595 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306619 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306642 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306687 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306721 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.306805 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.307436 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.308928 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.309056 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.309367 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.310081 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.313820 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.321680 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.328978 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.330234 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.333666 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwjb8\" (UniqueName: \"kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.342028 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:18 crc kubenswrapper[5002]: I1014 08:48:18.438266 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 14 08:48:19 crc kubenswrapper[5002]: I1014 08:48:19.002106 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Oct 14 08:48:19 crc kubenswrapper[5002]: I1014 08:48:19.689972 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"ec2597cf-678d-44a5-a3e3-41be976a3dc6","Type":"ContainerStarted","Data":"06ffbfddb91b06b31050e857f922d727725c302654a26a31cab154841fb669cb"} Oct 14 08:48:46 crc kubenswrapper[5002]: E1014 08:48:46.709131 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 14 08:48:46 crc kubenswrapper[5002]: E1014 08:48:46.709960 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zwjb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(ec2597cf-678d-44a5-a3e3-41be976a3dc6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 08:48:46 crc kubenswrapper[5002]: E1014 08:48:46.711350 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" Oct 14 08:48:46 crc kubenswrapper[5002]: E1014 08:48:46.979991 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" Oct 14 08:49:01 crc kubenswrapper[5002]: I1014 08:49:01.724216 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:49:04 crc kubenswrapper[5002]: I1014 08:49:04.176964 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"ec2597cf-678d-44a5-a3e3-41be976a3dc6","Type":"ContainerStarted","Data":"45f95251e4113681c6719ac32efb6accda6c204884b6beaa6ec54a9d93278500"} Oct 14 08:49:04 crc kubenswrapper[5002]: I1014 08:49:04.202789 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=3.641058496 podStartE2EDuration="47.202769987s" podCreationTimestamp="2025-10-14 08:48:17 +0000 UTC" firstStartedPulling="2025-10-14 08:48:19.015458128 +0000 UTC m=+3431.996697590" lastFinishedPulling="2025-10-14 08:49:02.577169629 +0000 UTC m=+3475.558409081" observedRunningTime="2025-10-14 08:49:04.200551198 +0000 UTC m=+3477.181790680" watchObservedRunningTime="2025-10-14 08:49:04.202769987 +0000 UTC m=+3477.184009439" Oct 14 08:49:09 crc kubenswrapper[5002]: I1014 08:49:09.218136 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:49:09 crc kubenswrapper[5002]: I1014 08:49:09.218889 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:49:39 crc kubenswrapper[5002]: I1014 08:49:39.218364 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:49:39 crc kubenswrapper[5002]: I1014 08:49:39.219181 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.218417 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.219323 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.219415 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.220693 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.220824 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" gracePeriod=600 Oct 14 08:50:09 crc kubenswrapper[5002]: E1014 08:50:09.346070 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.993515 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" exitCode=0 Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.993555 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207"} Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.993610 5002 scope.go:117] "RemoveContainer" containerID="41d268b5514016477348245720a6724ce2b5afb4e2706c65ca45dd6869f4a19a" Oct 14 08:50:09 crc kubenswrapper[5002]: I1014 08:50:09.994992 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:50:09 crc kubenswrapper[5002]: E1014 08:50:09.995505 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:50:18 crc kubenswrapper[5002]: I1014 08:50:18.140219 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7855f7b865-dtwlh" podUID="abe829e6-d42f-447f-bcb2-aad97c1e4f92" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 14 08:50:22 crc kubenswrapper[5002]: I1014 08:50:22.720123 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:50:22 crc kubenswrapper[5002]: E1014 08:50:22.721073 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:50:28 crc kubenswrapper[5002]: I1014 08:50:28.289569 5002 scope.go:117] "RemoveContainer" containerID="71cec4a0ac5eb7cd96f7f217f31c7f62dafa0bf16492820ad7af2e3ba551d454" Oct 14 08:50:28 crc kubenswrapper[5002]: I1014 08:50:28.326743 5002 scope.go:117] "RemoveContainer" containerID="a4e1b4f134a570a93b23a42a88a8016a0853080e9e2c12779e27ff3181e43e2b" Oct 14 08:50:36 crc kubenswrapper[5002]: I1014 08:50:36.721001 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:50:36 crc kubenswrapper[5002]: E1014 08:50:36.722083 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:50:47 crc kubenswrapper[5002]: I1014 08:50:47.732601 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:50:47 crc kubenswrapper[5002]: E1014 08:50:47.733395 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:51:00 crc kubenswrapper[5002]: I1014 08:51:00.720665 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:51:00 crc kubenswrapper[5002]: E1014 08:51:00.721766 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:51:12 crc kubenswrapper[5002]: I1014 08:51:12.720985 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:51:12 crc kubenswrapper[5002]: E1014 08:51:12.721869 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:51:27 crc kubenswrapper[5002]: I1014 08:51:27.749999 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:51:27 crc kubenswrapper[5002]: E1014 08:51:27.750870 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:51:41 crc kubenswrapper[5002]: I1014 08:51:41.722483 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:51:41 crc kubenswrapper[5002]: E1014 08:51:41.723666 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:51:52 crc kubenswrapper[5002]: I1014 08:51:52.721379 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:51:52 crc kubenswrapper[5002]: E1014 08:51:52.723007 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:52:05 crc kubenswrapper[5002]: I1014 08:52:05.721275 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:52:05 crc kubenswrapper[5002]: E1014 08:52:05.722485 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:52:20 crc kubenswrapper[5002]: I1014 08:52:20.721048 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:52:20 crc kubenswrapper[5002]: E1014 08:52:20.722232 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:52:32 crc kubenswrapper[5002]: I1014 08:52:32.720526 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:52:32 crc kubenswrapper[5002]: E1014 08:52:32.721203 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:52:45 crc kubenswrapper[5002]: I1014 08:52:45.720406 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:52:45 crc kubenswrapper[5002]: E1014 08:52:45.721660 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:52:57 crc kubenswrapper[5002]: I1014 08:52:57.748998 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:52:57 crc kubenswrapper[5002]: E1014 08:52:57.750894 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:53:12 crc kubenswrapper[5002]: I1014 08:53:12.720374 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:53:12 crc kubenswrapper[5002]: E1014 08:53:12.722258 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:53:23 crc kubenswrapper[5002]: I1014 08:53:23.721976 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:53:23 crc kubenswrapper[5002]: E1014 08:53:23.723126 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:53:25 crc kubenswrapper[5002]: I1014 08:53:25.044568 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-qqxjr"] Oct 14 08:53:25 crc kubenswrapper[5002]: I1014 08:53:25.051959 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-qqxjr"] Oct 14 08:53:25 crc kubenswrapper[5002]: I1014 08:53:25.730985 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="128b8127-99a6-4ffb-98fa-cff179c1ec85" path="/var/lib/kubelet/pods/128b8127-99a6-4ffb-98fa-cff179c1ec85/volumes" Oct 14 08:53:28 crc kubenswrapper[5002]: I1014 08:53:28.441236 5002 scope.go:117] "RemoveContainer" containerID="b793c0a550919208c0c7559888c2a04e211cbbdfa51bba65dfbbed6d5ee2b39f" Oct 14 08:53:34 crc kubenswrapper[5002]: I1014 08:53:34.721572 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:53:34 crc kubenswrapper[5002]: E1014 08:53:34.722246 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:53:35 crc kubenswrapper[5002]: I1014 08:53:35.057525 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-9b88-account-create-n5r4b"] Oct 14 08:53:35 crc kubenswrapper[5002]: I1014 08:53:35.064644 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-9b88-account-create-n5r4b"] Oct 14 08:53:35 crc kubenswrapper[5002]: I1014 08:53:35.736035 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="569d4442-3987-42ba-b796-286146ce2cdc" path="/var/lib/kubelet/pods/569d4442-3987-42ba-b796-286146ce2cdc/volumes" Oct 14 08:53:45 crc kubenswrapper[5002]: I1014 08:53:45.720431 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:53:45 crc kubenswrapper[5002]: E1014 08:53:45.721237 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:53:55 crc kubenswrapper[5002]: I1014 08:53:55.031388 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-ddkcm"] Oct 14 08:53:55 crc kubenswrapper[5002]: I1014 08:53:55.040794 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-ddkcm"] Oct 14 08:53:55 crc kubenswrapper[5002]: I1014 08:53:55.737613 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="431c8c01-259f-4ed7-8244-f16a549042ef" path="/var/lib/kubelet/pods/431c8c01-259f-4ed7-8244-f16a549042ef/volumes" Oct 14 08:53:59 crc kubenswrapper[5002]: I1014 08:53:59.721088 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:53:59 crc kubenswrapper[5002]: E1014 08:53:59.722008 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:54:13 crc kubenswrapper[5002]: I1014 08:54:13.725929 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:54:13 crc kubenswrapper[5002]: E1014 08:54:13.728572 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:54:25 crc kubenswrapper[5002]: I1014 08:54:25.721718 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:54:25 crc kubenswrapper[5002]: E1014 08:54:25.722933 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:54:28 crc kubenswrapper[5002]: I1014 08:54:28.512318 5002 scope.go:117] "RemoveContainer" containerID="f4774067f075249866c85c6b565ab3c2fd9a387c84e9810accf0a1e09da08a54" Oct 14 08:54:28 crc kubenswrapper[5002]: I1014 08:54:28.989822 5002 scope.go:117] "RemoveContainer" containerID="3a1f10ca2b6516e7a18ac25d93be6b9cb1770d1fe2c0c7cb124f8f64bb6530e2" Oct 14 08:54:37 crc kubenswrapper[5002]: I1014 08:54:37.726845 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:54:37 crc kubenswrapper[5002]: E1014 08:54:37.727549 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:54:52 crc kubenswrapper[5002]: I1014 08:54:52.721025 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:54:52 crc kubenswrapper[5002]: E1014 08:54:52.722008 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:55:07 crc kubenswrapper[5002]: I1014 08:55:07.725988 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:55:07 crc kubenswrapper[5002]: E1014 08:55:07.726815 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 08:55:22 crc kubenswrapper[5002]: I1014 08:55:22.721148 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:55:23 crc kubenswrapper[5002]: I1014 08:55:23.406236 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274"} Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.668573 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.671495 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.688185 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.778915 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.779036 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zkzr\" (UniqueName: \"kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.779069 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.881460 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.881539 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zkzr\" (UniqueName: \"kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.881562 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.881991 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.882437 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:10 crc kubenswrapper[5002]: I1014 08:57:10.901036 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zkzr\" (UniqueName: \"kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr\") pod \"redhat-operators-b6mfx\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:11 crc kubenswrapper[5002]: I1014 08:57:11.005081 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:11 crc kubenswrapper[5002]: I1014 08:57:11.535228 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:12 crc kubenswrapper[5002]: I1014 08:57:12.476591 5002 generic.go:334] "Generic (PLEG): container finished" podID="8dbee846-defb-4db6-a05b-254689858ff8" containerID="baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f" exitCode=0 Oct 14 08:57:12 crc kubenswrapper[5002]: I1014 08:57:12.476782 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerDied","Data":"baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f"} Oct 14 08:57:12 crc kubenswrapper[5002]: I1014 08:57:12.477251 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerStarted","Data":"fa8db6f0cce4923b26adce458aaeee0cb90ea71136b8bb8eead552ee8342553c"} Oct 14 08:57:12 crc kubenswrapper[5002]: I1014 08:57:12.483254 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:57:14 crc kubenswrapper[5002]: I1014 08:57:14.518208 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerStarted","Data":"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031"} Oct 14 08:57:15 crc kubenswrapper[5002]: I1014 08:57:15.527175 5002 generic.go:334] "Generic (PLEG): container finished" podID="8dbee846-defb-4db6-a05b-254689858ff8" containerID="476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031" exitCode=0 Oct 14 08:57:15 crc kubenswrapper[5002]: I1014 08:57:15.527339 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerDied","Data":"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031"} Oct 14 08:57:16 crc kubenswrapper[5002]: I1014 08:57:16.545598 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerStarted","Data":"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db"} Oct 14 08:57:16 crc kubenswrapper[5002]: I1014 08:57:16.575336 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b6mfx" podStartSLOduration=3.124590184 podStartE2EDuration="6.575314958s" podCreationTimestamp="2025-10-14 08:57:10 +0000 UTC" firstStartedPulling="2025-10-14 08:57:12.482873467 +0000 UTC m=+3965.464112939" lastFinishedPulling="2025-10-14 08:57:15.933598251 +0000 UTC m=+3968.914837713" observedRunningTime="2025-10-14 08:57:16.573628054 +0000 UTC m=+3969.554867516" watchObservedRunningTime="2025-10-14 08:57:16.575314958 +0000 UTC m=+3969.556554420" Oct 14 08:57:21 crc kubenswrapper[5002]: I1014 08:57:21.005912 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:21 crc kubenswrapper[5002]: I1014 08:57:21.006520 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:22 crc kubenswrapper[5002]: I1014 08:57:22.058905 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b6mfx" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="registry-server" probeResult="failure" output=< Oct 14 08:57:22 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 08:57:22 crc kubenswrapper[5002]: > Oct 14 08:57:31 crc kubenswrapper[5002]: I1014 08:57:31.059958 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:31 crc kubenswrapper[5002]: I1014 08:57:31.137247 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:31 crc kubenswrapper[5002]: I1014 08:57:31.299719 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:32 crc kubenswrapper[5002]: I1014 08:57:32.696447 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b6mfx" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="registry-server" containerID="cri-o://9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db" gracePeriod=2 Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.538296 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.693084 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities\") pod \"8dbee846-defb-4db6-a05b-254689858ff8\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.693232 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zkzr\" (UniqueName: \"kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr\") pod \"8dbee846-defb-4db6-a05b-254689858ff8\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.693321 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content\") pod \"8dbee846-defb-4db6-a05b-254689858ff8\" (UID: \"8dbee846-defb-4db6-a05b-254689858ff8\") " Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.694865 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities" (OuterVolumeSpecName: "utilities") pod "8dbee846-defb-4db6-a05b-254689858ff8" (UID: "8dbee846-defb-4db6-a05b-254689858ff8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.701111 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr" (OuterVolumeSpecName: "kube-api-access-2zkzr") pod "8dbee846-defb-4db6-a05b-254689858ff8" (UID: "8dbee846-defb-4db6-a05b-254689858ff8"). InnerVolumeSpecName "kube-api-access-2zkzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.707002 5002 generic.go:334] "Generic (PLEG): container finished" podID="8dbee846-defb-4db6-a05b-254689858ff8" containerID="9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db" exitCode=0 Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.707043 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerDied","Data":"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db"} Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.707070 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6mfx" event={"ID":"8dbee846-defb-4db6-a05b-254689858ff8","Type":"ContainerDied","Data":"fa8db6f0cce4923b26adce458aaeee0cb90ea71136b8bb8eead552ee8342553c"} Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.707073 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6mfx" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.707087 5002 scope.go:117] "RemoveContainer" containerID="9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.787661 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dbee846-defb-4db6-a05b-254689858ff8" (UID: "8dbee846-defb-4db6-a05b-254689858ff8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.788619 5002 scope.go:117] "RemoveContainer" containerID="476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.795640 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.795670 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zkzr\" (UniqueName: \"kubernetes.io/projected/8dbee846-defb-4db6-a05b-254689858ff8-kube-api-access-2zkzr\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.795686 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dbee846-defb-4db6-a05b-254689858ff8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.813687 5002 scope.go:117] "RemoveContainer" containerID="baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.875962 5002 scope.go:117] "RemoveContainer" containerID="9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db" Oct 14 08:57:33 crc kubenswrapper[5002]: E1014 08:57:33.876437 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db\": container with ID starting with 9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db not found: ID does not exist" containerID="9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.876496 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db"} err="failed to get container status \"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db\": rpc error: code = NotFound desc = could not find container \"9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db\": container with ID starting with 9cd56b7ae8ae9c75204e966c5a76933181e1ceb1e0895de3b5e8af9e9ea503db not found: ID does not exist" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.876522 5002 scope.go:117] "RemoveContainer" containerID="476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031" Oct 14 08:57:33 crc kubenswrapper[5002]: E1014 08:57:33.877312 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031\": container with ID starting with 476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031 not found: ID does not exist" containerID="476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.877345 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031"} err="failed to get container status \"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031\": rpc error: code = NotFound desc = could not find container \"476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031\": container with ID starting with 476289dd5ed38e43083d5da50fa803dcd6064b9eaee82288da90e3b145179031 not found: ID does not exist" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.877368 5002 scope.go:117] "RemoveContainer" containerID="baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f" Oct 14 08:57:33 crc kubenswrapper[5002]: E1014 08:57:33.877622 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f\": container with ID starting with baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f not found: ID does not exist" containerID="baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f" Oct 14 08:57:33 crc kubenswrapper[5002]: I1014 08:57:33.877645 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f"} err="failed to get container status \"baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f\": rpc error: code = NotFound desc = could not find container \"baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f\": container with ID starting with baa97def3dfdda870781aba507af7e83015164715ffc7479472bd2189bddfb3f not found: ID does not exist" Oct 14 08:57:34 crc kubenswrapper[5002]: I1014 08:57:34.039284 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:34 crc kubenswrapper[5002]: I1014 08:57:34.047765 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b6mfx"] Oct 14 08:57:35 crc kubenswrapper[5002]: I1014 08:57:35.732355 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dbee846-defb-4db6-a05b-254689858ff8" path="/var/lib/kubelet/pods/8dbee846-defb-4db6-a05b-254689858ff8/volumes" Oct 14 08:57:39 crc kubenswrapper[5002]: I1014 08:57:39.218826 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:57:39 crc kubenswrapper[5002]: I1014 08:57:39.219309 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.895796 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:08 crc kubenswrapper[5002]: E1014 08:58:08.896905 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="extract-content" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.896921 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="extract-content" Oct 14 08:58:08 crc kubenswrapper[5002]: E1014 08:58:08.896947 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="extract-utilities" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.896955 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="extract-utilities" Oct 14 08:58:08 crc kubenswrapper[5002]: E1014 08:58:08.896995 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="registry-server" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.897002 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="registry-server" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.897217 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbee846-defb-4db6-a05b-254689858ff8" containerName="registry-server" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.898890 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.914356 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.991522 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5t29\" (UniqueName: \"kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.991708 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:08 crc kubenswrapper[5002]: I1014 08:58:08.992161 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.092935 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.093038 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5t29\" (UniqueName: \"kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.093111 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.093680 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.093722 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.125379 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5t29\" (UniqueName: \"kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29\") pod \"community-operators-xfdb8\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.217871 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.217928 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.221859 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:09 crc kubenswrapper[5002]: I1014 08:58:09.787369 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:10 crc kubenswrapper[5002]: I1014 08:58:10.047212 5002 generic.go:334] "Generic (PLEG): container finished" podID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerID="e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9" exitCode=0 Oct 14 08:58:10 crc kubenswrapper[5002]: I1014 08:58:10.047323 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerDied","Data":"e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9"} Oct 14 08:58:10 crc kubenswrapper[5002]: I1014 08:58:10.048541 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerStarted","Data":"cd2a1af4635a831283d2625ddbb298e92623782426828096f2be1a4679a0e66f"} Oct 14 08:58:12 crc kubenswrapper[5002]: I1014 08:58:12.074054 5002 generic.go:334] "Generic (PLEG): container finished" podID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerID="8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0" exitCode=0 Oct 14 08:58:12 crc kubenswrapper[5002]: I1014 08:58:12.074700 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerDied","Data":"8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0"} Oct 14 08:58:14 crc kubenswrapper[5002]: I1014 08:58:14.096612 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerStarted","Data":"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781"} Oct 14 08:58:19 crc kubenswrapper[5002]: I1014 08:58:19.223284 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:19 crc kubenswrapper[5002]: I1014 08:58:19.224037 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:19 crc kubenswrapper[5002]: I1014 08:58:19.270152 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:19 crc kubenswrapper[5002]: I1014 08:58:19.294795 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xfdb8" podStartSLOduration=8.01932894 podStartE2EDuration="11.294777051s" podCreationTimestamp="2025-10-14 08:58:08 +0000 UTC" firstStartedPulling="2025-10-14 08:58:10.048868636 +0000 UTC m=+4023.030108088" lastFinishedPulling="2025-10-14 08:58:13.324316747 +0000 UTC m=+4026.305556199" observedRunningTime="2025-10-14 08:58:14.124273057 +0000 UTC m=+4027.105512519" watchObservedRunningTime="2025-10-14 08:58:19.294777051 +0000 UTC m=+4032.276016503" Oct 14 08:58:20 crc kubenswrapper[5002]: I1014 08:58:20.213481 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:20 crc kubenswrapper[5002]: I1014 08:58:20.271306 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:22 crc kubenswrapper[5002]: I1014 08:58:22.176223 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xfdb8" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="registry-server" containerID="cri-o://203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781" gracePeriod=2 Oct 14 08:58:22 crc kubenswrapper[5002]: I1014 08:58:22.818739 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.022283 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5t29\" (UniqueName: \"kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29\") pod \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.022769 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities\") pod \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.023107 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content\") pod \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\" (UID: \"0e6ad3a3-98f3-4acc-b48b-d32477eee201\") " Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.023699 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities" (OuterVolumeSpecName: "utilities") pod "0e6ad3a3-98f3-4acc-b48b-d32477eee201" (UID: "0e6ad3a3-98f3-4acc-b48b-d32477eee201"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.023813 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.038011 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29" (OuterVolumeSpecName: "kube-api-access-n5t29") pod "0e6ad3a3-98f3-4acc-b48b-d32477eee201" (UID: "0e6ad3a3-98f3-4acc-b48b-d32477eee201"). InnerVolumeSpecName "kube-api-access-n5t29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.072354 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e6ad3a3-98f3-4acc-b48b-d32477eee201" (UID: "0e6ad3a3-98f3-4acc-b48b-d32477eee201"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.125826 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e6ad3a3-98f3-4acc-b48b-d32477eee201-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.125908 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5t29\" (UniqueName: \"kubernetes.io/projected/0e6ad3a3-98f3-4acc-b48b-d32477eee201-kube-api-access-n5t29\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.189985 5002 generic.go:334] "Generic (PLEG): container finished" podID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerID="203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781" exitCode=0 Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.190038 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerDied","Data":"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781"} Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.190096 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfdb8" event={"ID":"0e6ad3a3-98f3-4acc-b48b-d32477eee201","Type":"ContainerDied","Data":"cd2a1af4635a831283d2625ddbb298e92623782426828096f2be1a4679a0e66f"} Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.190119 5002 scope.go:117] "RemoveContainer" containerID="203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.190109 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfdb8" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.231986 5002 scope.go:117] "RemoveContainer" containerID="8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.233813 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.242740 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xfdb8"] Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.253147 5002 scope.go:117] "RemoveContainer" containerID="e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.304710 5002 scope.go:117] "RemoveContainer" containerID="203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781" Oct 14 08:58:23 crc kubenswrapper[5002]: E1014 08:58:23.305895 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781\": container with ID starting with 203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781 not found: ID does not exist" containerID="203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.305925 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781"} err="failed to get container status \"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781\": rpc error: code = NotFound desc = could not find container \"203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781\": container with ID starting with 203e2d1b44c5d0d8ea9dd9f4cf2006f2cbecdf7eb16e8752dbcf104f4cd59781 not found: ID does not exist" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.305945 5002 scope.go:117] "RemoveContainer" containerID="8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0" Oct 14 08:58:23 crc kubenswrapper[5002]: E1014 08:58:23.307402 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0\": container with ID starting with 8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0 not found: ID does not exist" containerID="8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.307425 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0"} err="failed to get container status \"8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0\": rpc error: code = NotFound desc = could not find container \"8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0\": container with ID starting with 8be96482a904b41f7c0ce65194a2c17beaf1216111f780d6725df4c62bd142a0 not found: ID does not exist" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.307438 5002 scope.go:117] "RemoveContainer" containerID="e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9" Oct 14 08:58:23 crc kubenswrapper[5002]: E1014 08:58:23.307651 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9\": container with ID starting with e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9 not found: ID does not exist" containerID="e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.307670 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9"} err="failed to get container status \"e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9\": rpc error: code = NotFound desc = could not find container \"e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9\": container with ID starting with e78f35612d34165414436b00f72ae23c666a0e2d5420b6ee52585f134eed5ff9 not found: ID does not exist" Oct 14 08:58:23 crc kubenswrapper[5002]: I1014 08:58:23.733667 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" path="/var/lib/kubelet/pods/0e6ad3a3-98f3-4acc-b48b-d32477eee201/volumes" Oct 14 08:58:39 crc kubenswrapper[5002]: I1014 08:58:39.218782 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:58:39 crc kubenswrapper[5002]: I1014 08:58:39.219500 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:58:39 crc kubenswrapper[5002]: I1014 08:58:39.219569 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 08:58:39 crc kubenswrapper[5002]: I1014 08:58:39.220905 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:58:39 crc kubenswrapper[5002]: I1014 08:58:39.221060 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274" gracePeriod=600 Oct 14 08:58:40 crc kubenswrapper[5002]: I1014 08:58:40.368918 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274" exitCode=0 Oct 14 08:58:40 crc kubenswrapper[5002]: I1014 08:58:40.368960 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274"} Oct 14 08:58:40 crc kubenswrapper[5002]: I1014 08:58:40.369562 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca"} Oct 14 08:58:40 crc kubenswrapper[5002]: I1014 08:58:40.369585 5002 scope.go:117] "RemoveContainer" containerID="8d1074a99bb7f5ea3bb8346b591562a84969c8475457df184c1078e97bba5207" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.430980 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:58:59 crc kubenswrapper[5002]: E1014 08:58:59.432064 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="extract-utilities" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.432083 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="extract-utilities" Oct 14 08:58:59 crc kubenswrapper[5002]: E1014 08:58:59.432119 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="extract-content" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.432128 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="extract-content" Oct 14 08:58:59 crc kubenswrapper[5002]: E1014 08:58:59.432140 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="registry-server" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.432150 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="registry-server" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.432454 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e6ad3a3-98f3-4acc-b48b-d32477eee201" containerName="registry-server" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.434225 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.446675 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.483123 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.483375 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj7tf\" (UniqueName: \"kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.483416 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.585802 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj7tf\" (UniqueName: \"kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.585896 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.585951 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.586447 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.586537 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.627770 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj7tf\" (UniqueName: \"kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf\") pod \"certified-operators-w2md5\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:58:59 crc kubenswrapper[5002]: I1014 08:58:59.759060 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:00 crc kubenswrapper[5002]: I1014 08:59:00.295107 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:59:00 crc kubenswrapper[5002]: W1014 08:59:00.313988 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod343af690_a661_4c84_acb8_08e6dd0bd3ce.slice/crio-8704cc19d1c453752277f2c37d5f6a9b73b34b6be14491fede1f8273a6776605 WatchSource:0}: Error finding container 8704cc19d1c453752277f2c37d5f6a9b73b34b6be14491fede1f8273a6776605: Status 404 returned error can't find the container with id 8704cc19d1c453752277f2c37d5f6a9b73b34b6be14491fede1f8273a6776605 Oct 14 08:59:00 crc kubenswrapper[5002]: I1014 08:59:00.561120 5002 generic.go:334] "Generic (PLEG): container finished" podID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerID="9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218" exitCode=0 Oct 14 08:59:00 crc kubenswrapper[5002]: I1014 08:59:00.561166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerDied","Data":"9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218"} Oct 14 08:59:00 crc kubenswrapper[5002]: I1014 08:59:00.562306 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerStarted","Data":"8704cc19d1c453752277f2c37d5f6a9b73b34b6be14491fede1f8273a6776605"} Oct 14 08:59:01 crc kubenswrapper[5002]: I1014 08:59:01.572007 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerStarted","Data":"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981"} Oct 14 08:59:02 crc kubenswrapper[5002]: I1014 08:59:02.583026 5002 generic.go:334] "Generic (PLEG): container finished" podID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerID="28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981" exitCode=0 Oct 14 08:59:02 crc kubenswrapper[5002]: I1014 08:59:02.583109 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerDied","Data":"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981"} Oct 14 08:59:03 crc kubenswrapper[5002]: I1014 08:59:03.593651 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerStarted","Data":"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef"} Oct 14 08:59:03 crc kubenswrapper[5002]: I1014 08:59:03.613254 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w2md5" podStartSLOduration=2.045655904 podStartE2EDuration="4.613237578s" podCreationTimestamp="2025-10-14 08:58:59 +0000 UTC" firstStartedPulling="2025-10-14 08:59:00.563597488 +0000 UTC m=+4073.544836940" lastFinishedPulling="2025-10-14 08:59:03.131179152 +0000 UTC m=+4076.112418614" observedRunningTime="2025-10-14 08:59:03.611265396 +0000 UTC m=+4076.592504848" watchObservedRunningTime="2025-10-14 08:59:03.613237578 +0000 UTC m=+4076.594477030" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.209735 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.212047 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.234322 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.289204 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6brv\" (UniqueName: \"kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.289282 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.289465 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.392010 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6brv\" (UniqueName: \"kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.392105 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.392136 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.392600 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.392686 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.420013 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6brv\" (UniqueName: \"kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv\") pod \"redhat-marketplace-226km\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:04 crc kubenswrapper[5002]: I1014 08:59:04.532558 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:05 crc kubenswrapper[5002]: I1014 08:59:05.039979 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:06 crc kubenswrapper[5002]: I1014 08:59:06.623808 5002 generic.go:334] "Generic (PLEG): container finished" podID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerID="2bcf4adc7de760001476dcf953ba67e9e8b043ac504d9830512c1c608e83abc6" exitCode=0 Oct 14 08:59:06 crc kubenswrapper[5002]: I1014 08:59:06.623873 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerDied","Data":"2bcf4adc7de760001476dcf953ba67e9e8b043ac504d9830512c1c608e83abc6"} Oct 14 08:59:06 crc kubenswrapper[5002]: I1014 08:59:06.624302 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerStarted","Data":"9afe2db8a6b48697008d68357c7b91a17f649e37f995bee23d6ce3ec3f207ce3"} Oct 14 08:59:08 crc kubenswrapper[5002]: I1014 08:59:08.643220 5002 generic.go:334] "Generic (PLEG): container finished" podID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerID="8e2416040cfbc8b990407429ba2fa1e38ff32657b98e32fdb683f592e4db7cbf" exitCode=0 Oct 14 08:59:08 crc kubenswrapper[5002]: I1014 08:59:08.643419 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerDied","Data":"8e2416040cfbc8b990407429ba2fa1e38ff32657b98e32fdb683f592e4db7cbf"} Oct 14 08:59:09 crc kubenswrapper[5002]: I1014 08:59:09.658996 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerStarted","Data":"996160ba38cd30a73a1fc70eed2f5daa949fe85fd37367664aa66db5b0d60667"} Oct 14 08:59:09 crc kubenswrapper[5002]: I1014 08:59:09.691447 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-226km" podStartSLOduration=2.971283762 podStartE2EDuration="5.691427711s" podCreationTimestamp="2025-10-14 08:59:04 +0000 UTC" firstStartedPulling="2025-10-14 08:59:06.627268036 +0000 UTC m=+4079.608507528" lastFinishedPulling="2025-10-14 08:59:09.347412015 +0000 UTC m=+4082.328651477" observedRunningTime="2025-10-14 08:59:09.685823303 +0000 UTC m=+4082.667062795" watchObservedRunningTime="2025-10-14 08:59:09.691427711 +0000 UTC m=+4082.672667173" Oct 14 08:59:09 crc kubenswrapper[5002]: I1014 08:59:09.759265 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:09 crc kubenswrapper[5002]: I1014 08:59:09.759323 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:09 crc kubenswrapper[5002]: I1014 08:59:09.811125 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:10 crc kubenswrapper[5002]: I1014 08:59:10.727898 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:12 crc kubenswrapper[5002]: I1014 08:59:12.007498 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:59:12 crc kubenswrapper[5002]: I1014 08:59:12.685370 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w2md5" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="registry-server" containerID="cri-o://ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef" gracePeriod=2 Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.408376 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.579712 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities\") pod \"343af690-a661-4c84-acb8-08e6dd0bd3ce\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.579800 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj7tf\" (UniqueName: \"kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf\") pod \"343af690-a661-4c84-acb8-08e6dd0bd3ce\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.580086 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content\") pod \"343af690-a661-4c84-acb8-08e6dd0bd3ce\" (UID: \"343af690-a661-4c84-acb8-08e6dd0bd3ce\") " Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.580514 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities" (OuterVolumeSpecName: "utilities") pod "343af690-a661-4c84-acb8-08e6dd0bd3ce" (UID: "343af690-a661-4c84-acb8-08e6dd0bd3ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.580691 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.586720 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf" (OuterVolumeSpecName: "kube-api-access-bj7tf") pod "343af690-a661-4c84-acb8-08e6dd0bd3ce" (UID: "343af690-a661-4c84-acb8-08e6dd0bd3ce"). InnerVolumeSpecName "kube-api-access-bj7tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.624608 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "343af690-a661-4c84-acb8-08e6dd0bd3ce" (UID: "343af690-a661-4c84-acb8-08e6dd0bd3ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.682845 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj7tf\" (UniqueName: \"kubernetes.io/projected/343af690-a661-4c84-acb8-08e6dd0bd3ce-kube-api-access-bj7tf\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.682882 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343af690-a661-4c84-acb8-08e6dd0bd3ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.695700 5002 generic.go:334] "Generic (PLEG): container finished" podID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerID="ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef" exitCode=0 Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.695750 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerDied","Data":"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef"} Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.695787 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w2md5" event={"ID":"343af690-a661-4c84-acb8-08e6dd0bd3ce","Type":"ContainerDied","Data":"8704cc19d1c453752277f2c37d5f6a9b73b34b6be14491fede1f8273a6776605"} Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.695811 5002 scope.go:117] "RemoveContainer" containerID="ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.695998 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w2md5" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.733061 5002 scope.go:117] "RemoveContainer" containerID="28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.740596 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.748999 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w2md5"] Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.753891 5002 scope.go:117] "RemoveContainer" containerID="9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.795981 5002 scope.go:117] "RemoveContainer" containerID="ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef" Oct 14 08:59:13 crc kubenswrapper[5002]: E1014 08:59:13.798089 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef\": container with ID starting with ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef not found: ID does not exist" containerID="ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.798123 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef"} err="failed to get container status \"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef\": rpc error: code = NotFound desc = could not find container \"ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef\": container with ID starting with ef5257b43b281a652bbb9fea402bf4857f5ca840c6b9c8bd5ff1d97d02ffb4ef not found: ID does not exist" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.798143 5002 scope.go:117] "RemoveContainer" containerID="28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981" Oct 14 08:59:13 crc kubenswrapper[5002]: E1014 08:59:13.798569 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981\": container with ID starting with 28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981 not found: ID does not exist" containerID="28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.798588 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981"} err="failed to get container status \"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981\": rpc error: code = NotFound desc = could not find container \"28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981\": container with ID starting with 28a7cb881efc4cf8a02367d045da0ae242e6928fbe76d53f1f2c0a6445529981 not found: ID does not exist" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.798605 5002 scope.go:117] "RemoveContainer" containerID="9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218" Oct 14 08:59:13 crc kubenswrapper[5002]: E1014 08:59:13.799086 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218\": container with ID starting with 9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218 not found: ID does not exist" containerID="9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218" Oct 14 08:59:13 crc kubenswrapper[5002]: I1014 08:59:13.799120 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218"} err="failed to get container status \"9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218\": rpc error: code = NotFound desc = could not find container \"9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218\": container with ID starting with 9658412bcf5b295ac037b6ee5114dc9a45b675b44d9316b056bb0732bdd76218 not found: ID does not exist" Oct 14 08:59:14 crc kubenswrapper[5002]: I1014 08:59:14.533544 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:14 crc kubenswrapper[5002]: I1014 08:59:14.534056 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:14 crc kubenswrapper[5002]: I1014 08:59:14.618201 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:14 crc kubenswrapper[5002]: I1014 08:59:14.787555 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:15 crc kubenswrapper[5002]: I1014 08:59:15.744430 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" path="/var/lib/kubelet/pods/343af690-a661-4c84-acb8-08e6dd0bd3ce/volumes" Oct 14 08:59:16 crc kubenswrapper[5002]: I1014 08:59:16.411569 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:16 crc kubenswrapper[5002]: I1014 08:59:16.750034 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-226km" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="registry-server" containerID="cri-o://996160ba38cd30a73a1fc70eed2f5daa949fe85fd37367664aa66db5b0d60667" gracePeriod=2 Oct 14 08:59:17 crc kubenswrapper[5002]: I1014 08:59:17.761497 5002 generic.go:334] "Generic (PLEG): container finished" podID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerID="996160ba38cd30a73a1fc70eed2f5daa949fe85fd37367664aa66db5b0d60667" exitCode=0 Oct 14 08:59:17 crc kubenswrapper[5002]: I1014 08:59:17.761561 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerDied","Data":"996160ba38cd30a73a1fc70eed2f5daa949fe85fd37367664aa66db5b0d60667"} Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.107919 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.287456 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6brv\" (UniqueName: \"kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv\") pod \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.287534 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities\") pod \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.287641 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content\") pod \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\" (UID: \"b8e7a2d9-7d01-42be-bfc3-d0f82444b134\") " Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.289325 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities" (OuterVolumeSpecName: "utilities") pod "b8e7a2d9-7d01-42be-bfc3-d0f82444b134" (UID: "b8e7a2d9-7d01-42be-bfc3-d0f82444b134"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.293411 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv" (OuterVolumeSpecName: "kube-api-access-z6brv") pod "b8e7a2d9-7d01-42be-bfc3-d0f82444b134" (UID: "b8e7a2d9-7d01-42be-bfc3-d0f82444b134"). InnerVolumeSpecName "kube-api-access-z6brv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.316397 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8e7a2d9-7d01-42be-bfc3-d0f82444b134" (UID: "b8e7a2d9-7d01-42be-bfc3-d0f82444b134"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.391275 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6brv\" (UniqueName: \"kubernetes.io/projected/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-kube-api-access-z6brv\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.391322 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.391340 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8e7a2d9-7d01-42be-bfc3-d0f82444b134-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.774235 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-226km" event={"ID":"b8e7a2d9-7d01-42be-bfc3-d0f82444b134","Type":"ContainerDied","Data":"9afe2db8a6b48697008d68357c7b91a17f649e37f995bee23d6ce3ec3f207ce3"} Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.774304 5002 scope.go:117] "RemoveContainer" containerID="996160ba38cd30a73a1fc70eed2f5daa949fe85fd37367664aa66db5b0d60667" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.774501 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-226km" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.841118 5002 scope.go:117] "RemoveContainer" containerID="8e2416040cfbc8b990407429ba2fa1e38ff32657b98e32fdb683f592e4db7cbf" Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.851001 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:18 crc kubenswrapper[5002]: I1014 08:59:18.861967 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-226km"] Oct 14 08:59:19 crc kubenswrapper[5002]: I1014 08:59:19.275692 5002 scope.go:117] "RemoveContainer" containerID="2bcf4adc7de760001476dcf953ba67e9e8b043ac504d9830512c1c608e83abc6" Oct 14 08:59:19 crc kubenswrapper[5002]: I1014 08:59:19.733613 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" path="/var/lib/kubelet/pods/b8e7a2d9-7d01-42be-bfc3-d0f82444b134/volumes" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.164946 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd"] Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.167409 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="extract-utilities" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.167562 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="extract-utilities" Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.167688 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.167796 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.167962 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="extract-content" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.168075 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="extract-content" Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.168201 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="extract-content" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.168313 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="extract-content" Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.168435 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="extract-utilities" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.168550 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="extract-utilities" Oct 14 09:00:00 crc kubenswrapper[5002]: E1014 09:00:00.168675 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.168782 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.169307 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="343af690-a661-4c84-acb8-08e6dd0bd3ce" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.169455 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e7a2d9-7d01-42be-bfc3-d0f82444b134" containerName="registry-server" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.170531 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.172696 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.173062 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.181910 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd"] Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.304144 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msnlf\" (UniqueName: \"kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.304555 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.304597 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.405944 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msnlf\" (UniqueName: \"kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.406171 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.406227 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.407200 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.412580 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.423287 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msnlf\" (UniqueName: \"kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf\") pod \"collect-profiles-29340540-74twd\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.499182 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:00 crc kubenswrapper[5002]: I1014 09:00:00.954132 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd"] Oct 14 09:00:01 crc kubenswrapper[5002]: I1014 09:00:01.210919 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" event={"ID":"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576","Type":"ContainerStarted","Data":"87396fdb54dc15f9262c2ac19f25393dad96fa8208246678890ebc7865f1f694"} Oct 14 09:00:02 crc kubenswrapper[5002]: I1014 09:00:02.224451 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" event={"ID":"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576","Type":"ContainerStarted","Data":"d59821667e674e4349807a53c5eb7f3ebb2b80531840615c47dac5ad11573b84"} Oct 14 09:00:02 crc kubenswrapper[5002]: I1014 09:00:02.250723 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" podStartSLOduration=2.250702387 podStartE2EDuration="2.250702387s" podCreationTimestamp="2025-10-14 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:02.241774101 +0000 UTC m=+4135.223013573" watchObservedRunningTime="2025-10-14 09:00:02.250702387 +0000 UTC m=+4135.231941839" Oct 14 09:00:03 crc kubenswrapper[5002]: I1014 09:00:03.236630 5002 generic.go:334] "Generic (PLEG): container finished" podID="d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" containerID="d59821667e674e4349807a53c5eb7f3ebb2b80531840615c47dac5ad11573b84" exitCode=0 Oct 14 09:00:03 crc kubenswrapper[5002]: I1014 09:00:03.236752 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" event={"ID":"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576","Type":"ContainerDied","Data":"d59821667e674e4349807a53c5eb7f3ebb2b80531840615c47dac5ad11573b84"} Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.781407 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.899235 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume\") pod \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.899601 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume\") pod \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.899660 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msnlf\" (UniqueName: \"kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf\") pod \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\" (UID: \"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576\") " Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.899988 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume" (OuterVolumeSpecName: "config-volume") pod "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" (UID: "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.900166 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.909229 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" (UID: "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:04 crc kubenswrapper[5002]: I1014 09:00:04.910499 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf" (OuterVolumeSpecName: "kube-api-access-msnlf") pod "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" (UID: "d99e72de-1dbc-4ca2-8fd7-0f04e35d6576"). InnerVolumeSpecName "kube-api-access-msnlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.002190 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msnlf\" (UniqueName: \"kubernetes.io/projected/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-kube-api-access-msnlf\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.002233 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.254747 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" event={"ID":"d99e72de-1dbc-4ca2-8fd7-0f04e35d6576","Type":"ContainerDied","Data":"87396fdb54dc15f9262c2ac19f25393dad96fa8208246678890ebc7865f1f694"} Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.254782 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87396fdb54dc15f9262c2ac19f25393dad96fa8208246678890ebc7865f1f694" Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.254808 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd" Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.313218 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh"] Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.323390 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-25mdh"] Oct 14 09:00:05 crc kubenswrapper[5002]: I1014 09:00:05.730887 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4efd0fd5-8a8f-401f-a14f-ea0dee615274" path="/var/lib/kubelet/pods/4efd0fd5-8a8f-401f-a14f-ea0dee615274/volumes" Oct 14 09:00:29 crc kubenswrapper[5002]: I1014 09:00:29.379853 5002 scope.go:117] "RemoveContainer" containerID="f76d702845733377cb3c900878a6f406025144e810740b65c715deeb7c43e8d6" Oct 14 09:00:39 crc kubenswrapper[5002]: I1014 09:00:39.217962 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:00:39 crc kubenswrapper[5002]: I1014 09:00:39.218525 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.177363 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29340541-twn6l"] Oct 14 09:01:00 crc kubenswrapper[5002]: E1014 09:01:00.178370 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.178385 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.178626 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" containerName="collect-profiles" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.179461 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.213923 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-twn6l"] Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.300897 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5zzn\" (UniqueName: \"kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.301637 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.301815 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.301879 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.403498 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.403615 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.403639 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.403792 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5zzn\" (UniqueName: \"kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.422042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.422102 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.426251 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.429089 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5zzn\" (UniqueName: \"kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn\") pod \"keystone-cron-29340541-twn6l\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.503878 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:00 crc kubenswrapper[5002]: I1014 09:01:00.985834 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-twn6l"] Oct 14 09:01:01 crc kubenswrapper[5002]: I1014 09:01:01.789928 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-twn6l" event={"ID":"181a00bb-d0d0-483b-a51e-a84e308f06c8","Type":"ContainerStarted","Data":"b34a63def327a882ce6824f30530eb0698e075bc4fe028887246cdfa3560f716"} Oct 14 09:01:01 crc kubenswrapper[5002]: I1014 09:01:01.790219 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-twn6l" event={"ID":"181a00bb-d0d0-483b-a51e-a84e308f06c8","Type":"ContainerStarted","Data":"41fef5b4cb0fcb0270cc21f26f1d18d4d5d01d242bd4ca272b328f835d7325b3"} Oct 14 09:01:03 crc kubenswrapper[5002]: I1014 09:01:03.812163 5002 generic.go:334] "Generic (PLEG): container finished" podID="181a00bb-d0d0-483b-a51e-a84e308f06c8" containerID="b34a63def327a882ce6824f30530eb0698e075bc4fe028887246cdfa3560f716" exitCode=0 Oct 14 09:01:03 crc kubenswrapper[5002]: I1014 09:01:03.812245 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-twn6l" event={"ID":"181a00bb-d0d0-483b-a51e-a84e308f06c8","Type":"ContainerDied","Data":"b34a63def327a882ce6824f30530eb0698e075bc4fe028887246cdfa3560f716"} Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.330984 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.405264 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5zzn\" (UniqueName: \"kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn\") pod \"181a00bb-d0d0-483b-a51e-a84e308f06c8\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.405431 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle\") pod \"181a00bb-d0d0-483b-a51e-a84e308f06c8\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.405482 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data\") pod \"181a00bb-d0d0-483b-a51e-a84e308f06c8\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.405541 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys\") pod \"181a00bb-d0d0-483b-a51e-a84e308f06c8\" (UID: \"181a00bb-d0d0-483b-a51e-a84e308f06c8\") " Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.411805 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn" (OuterVolumeSpecName: "kube-api-access-x5zzn") pod "181a00bb-d0d0-483b-a51e-a84e308f06c8" (UID: "181a00bb-d0d0-483b-a51e-a84e308f06c8"). InnerVolumeSpecName "kube-api-access-x5zzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.417004 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "181a00bb-d0d0-483b-a51e-a84e308f06c8" (UID: "181a00bb-d0d0-483b-a51e-a84e308f06c8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.479792 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "181a00bb-d0d0-483b-a51e-a84e308f06c8" (UID: "181a00bb-d0d0-483b-a51e-a84e308f06c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.493519 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data" (OuterVolumeSpecName: "config-data") pod "181a00bb-d0d0-483b-a51e-a84e308f06c8" (UID: "181a00bb-d0d0-483b-a51e-a84e308f06c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.508402 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5zzn\" (UniqueName: \"kubernetes.io/projected/181a00bb-d0d0-483b-a51e-a84e308f06c8-kube-api-access-x5zzn\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.508437 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.508448 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.508456 5002 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/181a00bb-d0d0-483b-a51e-a84e308f06c8-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.834039 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-twn6l" event={"ID":"181a00bb-d0d0-483b-a51e-a84e308f06c8","Type":"ContainerDied","Data":"41fef5b4cb0fcb0270cc21f26f1d18d4d5d01d242bd4ca272b328f835d7325b3"} Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.834095 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41fef5b4cb0fcb0270cc21f26f1d18d4d5d01d242bd4ca272b328f835d7325b3" Oct 14 09:01:05 crc kubenswrapper[5002]: I1014 09:01:05.834138 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-twn6l" Oct 14 09:01:09 crc kubenswrapper[5002]: I1014 09:01:09.217946 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:01:09 crc kubenswrapper[5002]: I1014 09:01:09.218557 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:01:39 crc kubenswrapper[5002]: I1014 09:01:39.218025 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:01:39 crc kubenswrapper[5002]: I1014 09:01:39.218553 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:01:39 crc kubenswrapper[5002]: I1014 09:01:39.218600 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:01:39 crc kubenswrapper[5002]: I1014 09:01:39.219452 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:01:39 crc kubenswrapper[5002]: I1014 09:01:39.219521 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" gracePeriod=600 Oct 14 09:01:39 crc kubenswrapper[5002]: E1014 09:01:39.336796 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:01:40 crc kubenswrapper[5002]: I1014 09:01:40.161896 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" exitCode=0 Oct 14 09:01:40 crc kubenswrapper[5002]: I1014 09:01:40.161988 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca"} Oct 14 09:01:40 crc kubenswrapper[5002]: I1014 09:01:40.162217 5002 scope.go:117] "RemoveContainer" containerID="e3013caa4f7c6c38f1b592b0af18ee38dcc66217f07304affcc3a5f9fcc3c274" Oct 14 09:01:40 crc kubenswrapper[5002]: I1014 09:01:40.163014 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:01:40 crc kubenswrapper[5002]: E1014 09:01:40.163396 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:01:52 crc kubenswrapper[5002]: I1014 09:01:52.720654 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:01:52 crc kubenswrapper[5002]: E1014 09:01:52.721443 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:02:07 crc kubenswrapper[5002]: I1014 09:02:07.727284 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:02:07 crc kubenswrapper[5002]: E1014 09:02:07.728071 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:02:18 crc kubenswrapper[5002]: I1014 09:02:18.720879 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:02:18 crc kubenswrapper[5002]: E1014 09:02:18.721764 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:02:29 crc kubenswrapper[5002]: I1014 09:02:29.720609 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:02:29 crc kubenswrapper[5002]: E1014 09:02:29.722776 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:02:44 crc kubenswrapper[5002]: I1014 09:02:44.720942 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:02:44 crc kubenswrapper[5002]: E1014 09:02:44.721931 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:02:58 crc kubenswrapper[5002]: I1014 09:02:58.720018 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:02:58 crc kubenswrapper[5002]: E1014 09:02:58.721186 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:03:11 crc kubenswrapper[5002]: I1014 09:03:11.720871 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:03:11 crc kubenswrapper[5002]: E1014 09:03:11.721689 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:03:25 crc kubenswrapper[5002]: I1014 09:03:25.721187 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:03:25 crc kubenswrapper[5002]: E1014 09:03:25.722031 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:03:38 crc kubenswrapper[5002]: I1014 09:03:38.721166 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:03:38 crc kubenswrapper[5002]: E1014 09:03:38.722461 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:03:49 crc kubenswrapper[5002]: I1014 09:03:49.720225 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:03:49 crc kubenswrapper[5002]: E1014 09:03:49.721995 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:04:00 crc kubenswrapper[5002]: I1014 09:04:00.721318 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:04:00 crc kubenswrapper[5002]: E1014 09:04:00.722482 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:04:12 crc kubenswrapper[5002]: I1014 09:04:12.721329 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:04:12 crc kubenswrapper[5002]: E1014 09:04:12.722120 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:04:26 crc kubenswrapper[5002]: I1014 09:04:26.720310 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:04:26 crc kubenswrapper[5002]: E1014 09:04:26.721287 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:04:39 crc kubenswrapper[5002]: I1014 09:04:39.721910 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:04:39 crc kubenswrapper[5002]: E1014 09:04:39.725465 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:04:53 crc kubenswrapper[5002]: I1014 09:04:53.721401 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:04:53 crc kubenswrapper[5002]: E1014 09:04:53.722505 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:05:08 crc kubenswrapper[5002]: I1014 09:05:08.721394 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:05:08 crc kubenswrapper[5002]: E1014 09:05:08.722124 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:05:22 crc kubenswrapper[5002]: I1014 09:05:22.720717 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:05:22 crc kubenswrapper[5002]: E1014 09:05:22.721578 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:05:33 crc kubenswrapper[5002]: I1014 09:05:33.720963 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:05:33 crc kubenswrapper[5002]: E1014 09:05:33.722191 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:05:48 crc kubenswrapper[5002]: I1014 09:05:48.720746 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:05:48 crc kubenswrapper[5002]: E1014 09:05:48.721799 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:06:03 crc kubenswrapper[5002]: I1014 09:06:03.720714 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:06:03 crc kubenswrapper[5002]: E1014 09:06:03.721497 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:06:16 crc kubenswrapper[5002]: I1014 09:06:16.721238 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:06:16 crc kubenswrapper[5002]: E1014 09:06:16.722082 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:06:30 crc kubenswrapper[5002]: I1014 09:06:30.721792 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:06:30 crc kubenswrapper[5002]: E1014 09:06:30.722740 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:06:45 crc kubenswrapper[5002]: I1014 09:06:45.721605 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:06:46 crc kubenswrapper[5002]: I1014 09:06:46.185260 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b"} Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.721493 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:07:48 crc kubenswrapper[5002]: E1014 09:07:48.722900 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="181a00bb-d0d0-483b-a51e-a84e308f06c8" containerName="keystone-cron" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.722932 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="181a00bb-d0d0-483b-a51e-a84e308f06c8" containerName="keystone-cron" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.723353 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="181a00bb-d0d0-483b-a51e-a84e308f06c8" containerName="keystone-cron" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.727051 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.733159 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.868424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.868511 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8kc\" (UniqueName: \"kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.868996 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.970612 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.970924 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8kc\" (UniqueName: \"kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.971232 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.971441 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.971817 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:48 crc kubenswrapper[5002]: I1014 09:07:48.990643 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8kc\" (UniqueName: \"kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc\") pod \"redhat-operators-vhzpq\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:49 crc kubenswrapper[5002]: I1014 09:07:49.065698 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:49 crc kubenswrapper[5002]: I1014 09:07:49.612620 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:07:49 crc kubenswrapper[5002]: I1014 09:07:49.872467 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerStarted","Data":"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c"} Oct 14 09:07:49 crc kubenswrapper[5002]: I1014 09:07:49.872509 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerStarted","Data":"4dba6c97ba4d824c68f61a2a8f4f1a7ecedd83b83e6592f94c98b98528ae2ece"} Oct 14 09:07:50 crc kubenswrapper[5002]: I1014 09:07:50.884136 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerDied","Data":"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c"} Oct 14 09:07:50 crc kubenswrapper[5002]: I1014 09:07:50.883950 5002 generic.go:334] "Generic (PLEG): container finished" podID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerID="f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c" exitCode=0 Oct 14 09:07:50 crc kubenswrapper[5002]: I1014 09:07:50.887151 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:07:52 crc kubenswrapper[5002]: I1014 09:07:52.910390 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerStarted","Data":"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad"} Oct 14 09:07:53 crc kubenswrapper[5002]: I1014 09:07:53.925346 5002 generic.go:334] "Generic (PLEG): container finished" podID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerID="9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad" exitCode=0 Oct 14 09:07:53 crc kubenswrapper[5002]: I1014 09:07:53.926061 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerDied","Data":"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad"} Oct 14 09:07:54 crc kubenswrapper[5002]: I1014 09:07:54.942660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerStarted","Data":"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a"} Oct 14 09:07:59 crc kubenswrapper[5002]: I1014 09:07:59.066406 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:07:59 crc kubenswrapper[5002]: I1014 09:07:59.066870 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:08:00 crc kubenswrapper[5002]: I1014 09:08:00.134310 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vhzpq" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="registry-server" probeResult="failure" output=< Oct 14 09:08:00 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:08:00 crc kubenswrapper[5002]: > Oct 14 09:08:09 crc kubenswrapper[5002]: I1014 09:08:09.159671 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:08:09 crc kubenswrapper[5002]: I1014 09:08:09.207134 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vhzpq" podStartSLOduration=17.683367392 podStartE2EDuration="21.207109056s" podCreationTimestamp="2025-10-14 09:07:48 +0000 UTC" firstStartedPulling="2025-10-14 09:07:50.886807883 +0000 UTC m=+4603.868047345" lastFinishedPulling="2025-10-14 09:07:54.410549527 +0000 UTC m=+4607.391789009" observedRunningTime="2025-10-14 09:07:54.977797395 +0000 UTC m=+4607.959036887" watchObservedRunningTime="2025-10-14 09:08:09.207109056 +0000 UTC m=+4622.188348538" Oct 14 09:08:09 crc kubenswrapper[5002]: I1014 09:08:09.253310 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:08:09 crc kubenswrapper[5002]: I1014 09:08:09.429349 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.104714 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vhzpq" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="registry-server" containerID="cri-o://82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a" gracePeriod=2 Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.920399 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.946914 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities\") pod \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.946991 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl8kc\" (UniqueName: \"kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc\") pod \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.947207 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content\") pod \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\" (UID: \"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7\") " Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.947681 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities" (OuterVolumeSpecName: "utilities") pod "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" (UID: "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.948111 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:11 crc kubenswrapper[5002]: I1014 09:08:11.954034 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc" (OuterVolumeSpecName: "kube-api-access-bl8kc") pod "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" (UID: "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7"). InnerVolumeSpecName "kube-api-access-bl8kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.050220 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl8kc\" (UniqueName: \"kubernetes.io/projected/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-kube-api-access-bl8kc\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.066499 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" (UID: "c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.114601 5002 generic.go:334] "Generic (PLEG): container finished" podID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerID="82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a" exitCode=0 Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.114637 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerDied","Data":"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a"} Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.114685 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhzpq" event={"ID":"c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7","Type":"ContainerDied","Data":"4dba6c97ba4d824c68f61a2a8f4f1a7ecedd83b83e6592f94c98b98528ae2ece"} Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.114685 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhzpq" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.114701 5002 scope.go:117] "RemoveContainer" containerID="82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.141183 5002 scope.go:117] "RemoveContainer" containerID="9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.152599 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.153300 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.160236 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vhzpq"] Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.174223 5002 scope.go:117] "RemoveContainer" containerID="f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.205191 5002 scope.go:117] "RemoveContainer" containerID="82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a" Oct 14 09:08:12 crc kubenswrapper[5002]: E1014 09:08:12.205750 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a\": container with ID starting with 82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a not found: ID does not exist" containerID="82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.205896 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a"} err="failed to get container status \"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a\": rpc error: code = NotFound desc = could not find container \"82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a\": container with ID starting with 82bd2eed5cf8100bb6098cf7009885c2b74d15f34056e55a8f7801dbd895dd3a not found: ID does not exist" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.206000 5002 scope.go:117] "RemoveContainer" containerID="9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad" Oct 14 09:08:12 crc kubenswrapper[5002]: E1014 09:08:12.206613 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad\": container with ID starting with 9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad not found: ID does not exist" containerID="9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.206759 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad"} err="failed to get container status \"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad\": rpc error: code = NotFound desc = could not find container \"9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad\": container with ID starting with 9e440b2647bac3b20101f1d8253a2392bdb0025492512a25e247b673d2653aad not found: ID does not exist" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.206924 5002 scope.go:117] "RemoveContainer" containerID="f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c" Oct 14 09:08:12 crc kubenswrapper[5002]: E1014 09:08:12.207309 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c\": container with ID starting with f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c not found: ID does not exist" containerID="f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c" Oct 14 09:08:12 crc kubenswrapper[5002]: I1014 09:08:12.207460 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c"} err="failed to get container status \"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c\": rpc error: code = NotFound desc = could not find container \"f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c\": container with ID starting with f1f80c39be10d67da1ac27e8226887f40482b0115c6911a583db384219f2a00c not found: ID does not exist" Oct 14 09:08:13 crc kubenswrapper[5002]: I1014 09:08:13.764158 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" path="/var/lib/kubelet/pods/c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7/volumes" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.780098 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:27 crc kubenswrapper[5002]: E1014 09:08:27.782446 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="extract-content" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.782460 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="extract-content" Oct 14 09:08:27 crc kubenswrapper[5002]: E1014 09:08:27.782474 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="registry-server" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.782480 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="registry-server" Oct 14 09:08:27 crc kubenswrapper[5002]: E1014 09:08:27.782500 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="extract-utilities" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.782506 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="extract-utilities" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.782799 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2bb29c9-aeb3-4d2e-8ff1-807f5ea135f7" containerName="registry-server" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.784316 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.795550 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.986280 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.986483 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxcpf\" (UniqueName: \"kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:27 crc kubenswrapper[5002]: I1014 09:08:27.986512 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.087795 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxcpf\" (UniqueName: \"kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.087861 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.087942 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.088552 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.088832 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.112936 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxcpf\" (UniqueName: \"kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf\") pod \"community-operators-npfzv\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.409331 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:28 crc kubenswrapper[5002]: I1014 09:08:28.854256 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:29 crc kubenswrapper[5002]: I1014 09:08:29.285248 5002 generic.go:334] "Generic (PLEG): container finished" podID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerID="d8148bfe36b6be6cd76ae6efd19a4491e6345c35d9b567c5a991b3055eea0517" exitCode=0 Oct 14 09:08:29 crc kubenswrapper[5002]: I1014 09:08:29.285374 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerDied","Data":"d8148bfe36b6be6cd76ae6efd19a4491e6345c35d9b567c5a991b3055eea0517"} Oct 14 09:08:29 crc kubenswrapper[5002]: I1014 09:08:29.285583 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerStarted","Data":"79938a9d58c6193cb3e03d8ead6c5f39251230a3beb82dcfe198e55c1135f299"} Oct 14 09:08:31 crc kubenswrapper[5002]: I1014 09:08:31.303921 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerStarted","Data":"ab529e3079186dd26ba252a6f996f8528a977e88c951945a29cde9bf8ec6a016"} Oct 14 09:08:32 crc kubenswrapper[5002]: I1014 09:08:32.313719 5002 generic.go:334] "Generic (PLEG): container finished" podID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerID="ab529e3079186dd26ba252a6f996f8528a977e88c951945a29cde9bf8ec6a016" exitCode=0 Oct 14 09:08:32 crc kubenswrapper[5002]: I1014 09:08:32.313863 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerDied","Data":"ab529e3079186dd26ba252a6f996f8528a977e88c951945a29cde9bf8ec6a016"} Oct 14 09:08:34 crc kubenswrapper[5002]: I1014 09:08:34.341989 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerStarted","Data":"9113302b2fa99780232205eaeece2eeab2c1b2b5305bfb6cbf01624cc2a86210"} Oct 14 09:08:34 crc kubenswrapper[5002]: I1014 09:08:34.371456 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-npfzv" podStartSLOduration=3.37200636 podStartE2EDuration="7.37143063s" podCreationTimestamp="2025-10-14 09:08:27 +0000 UTC" firstStartedPulling="2025-10-14 09:08:29.289178161 +0000 UTC m=+4642.270417613" lastFinishedPulling="2025-10-14 09:08:33.288602431 +0000 UTC m=+4646.269841883" observedRunningTime="2025-10-14 09:08:34.361113638 +0000 UTC m=+4647.342353120" watchObservedRunningTime="2025-10-14 09:08:34.37143063 +0000 UTC m=+4647.352670122" Oct 14 09:08:38 crc kubenswrapper[5002]: I1014 09:08:38.409988 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:38 crc kubenswrapper[5002]: I1014 09:08:38.410569 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:38 crc kubenswrapper[5002]: I1014 09:08:38.470689 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:39 crc kubenswrapper[5002]: I1014 09:08:39.457854 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:39 crc kubenswrapper[5002]: I1014 09:08:39.520713 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:41 crc kubenswrapper[5002]: I1014 09:08:41.408580 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-npfzv" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="registry-server" containerID="cri-o://9113302b2fa99780232205eaeece2eeab2c1b2b5305bfb6cbf01624cc2a86210" gracePeriod=2 Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.422017 5002 generic.go:334] "Generic (PLEG): container finished" podID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerID="9113302b2fa99780232205eaeece2eeab2c1b2b5305bfb6cbf01624cc2a86210" exitCode=0 Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.422298 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerDied","Data":"9113302b2fa99780232205eaeece2eeab2c1b2b5305bfb6cbf01624cc2a86210"} Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.774030 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.901085 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities\") pod \"e6f5084c-cb14-455f-bee8-ec461101f69f\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.902505 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities" (OuterVolumeSpecName: "utilities") pod "e6f5084c-cb14-455f-bee8-ec461101f69f" (UID: "e6f5084c-cb14-455f-bee8-ec461101f69f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.903607 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxcpf\" (UniqueName: \"kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf\") pod \"e6f5084c-cb14-455f-bee8-ec461101f69f\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.903719 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content\") pod \"e6f5084c-cb14-455f-bee8-ec461101f69f\" (UID: \"e6f5084c-cb14-455f-bee8-ec461101f69f\") " Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.905039 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.910075 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf" (OuterVolumeSpecName: "kube-api-access-vxcpf") pod "e6f5084c-cb14-455f-bee8-ec461101f69f" (UID: "e6f5084c-cb14-455f-bee8-ec461101f69f"). InnerVolumeSpecName "kube-api-access-vxcpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:42 crc kubenswrapper[5002]: I1014 09:08:42.951763 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6f5084c-cb14-455f-bee8-ec461101f69f" (UID: "e6f5084c-cb14-455f-bee8-ec461101f69f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.006985 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxcpf\" (UniqueName: \"kubernetes.io/projected/e6f5084c-cb14-455f-bee8-ec461101f69f-kube-api-access-vxcpf\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.007020 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6f5084c-cb14-455f-bee8-ec461101f69f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.436058 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-npfzv" event={"ID":"e6f5084c-cb14-455f-bee8-ec461101f69f","Type":"ContainerDied","Data":"79938a9d58c6193cb3e03d8ead6c5f39251230a3beb82dcfe198e55c1135f299"} Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.436125 5002 scope.go:117] "RemoveContainer" containerID="9113302b2fa99780232205eaeece2eeab2c1b2b5305bfb6cbf01624cc2a86210" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.436334 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-npfzv" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.488772 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.499647 5002 scope.go:117] "RemoveContainer" containerID="ab529e3079186dd26ba252a6f996f8528a977e88c951945a29cde9bf8ec6a016" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.501821 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-npfzv"] Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.732695 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" path="/var/lib/kubelet/pods/e6f5084c-cb14-455f-bee8-ec461101f69f/volumes" Oct 14 09:08:43 crc kubenswrapper[5002]: I1014 09:08:43.871285 5002 scope.go:117] "RemoveContainer" containerID="d8148bfe36b6be6cd76ae6efd19a4491e6345c35d9b567c5a991b3055eea0517" Oct 14 09:09:09 crc kubenswrapper[5002]: I1014 09:09:09.218540 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:09:09 crc kubenswrapper[5002]: I1014 09:09:09.219239 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.204779 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:27 crc kubenswrapper[5002]: E1014 09:09:27.205886 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="extract-utilities" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.205900 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="extract-utilities" Oct 14 09:09:27 crc kubenswrapper[5002]: E1014 09:09:27.205917 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="extract-content" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.205924 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="extract-content" Oct 14 09:09:27 crc kubenswrapper[5002]: E1014 09:09:27.205949 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="registry-server" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.205955 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="registry-server" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.208658 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6f5084c-cb14-455f-bee8-ec461101f69f" containerName="registry-server" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.210654 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.228286 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.302337 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mbql\" (UniqueName: \"kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.304057 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.305077 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.406754 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mbql\" (UniqueName: \"kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.406816 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.406955 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.407662 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.407689 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.447169 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mbql\" (UniqueName: \"kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql\") pod \"redhat-marketplace-gq2pn\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:27 crc kubenswrapper[5002]: I1014 09:09:27.531873 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.048050 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.594334 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.596552 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.610636 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.628027 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.628242 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.628387 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzx4n\" (UniqueName: \"kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.730607 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.730812 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzx4n\" (UniqueName: \"kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.730963 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.731645 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.733941 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.758259 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzx4n\" (UniqueName: \"kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n\") pod \"certified-operators-wk5x7\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.853284 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:28 crc kubenswrapper[5002]: I1014 09:09:28.889903 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerStarted","Data":"c7ed5204751e7956d0b6829defecff6cbbec0ec13935381482162ca373d813f9"} Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.441126 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:29 crc kubenswrapper[5002]: W1014 09:09:29.460282 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod763f3c28_d977_4ebc_81a1_e774138472c3.slice/crio-88dc6550a12932c59334207c9027ee76788b100378df4b83fdf05a8ed60d6d9d WatchSource:0}: Error finding container 88dc6550a12932c59334207c9027ee76788b100378df4b83fdf05a8ed60d6d9d: Status 404 returned error can't find the container with id 88dc6550a12932c59334207c9027ee76788b100378df4b83fdf05a8ed60d6d9d Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.901167 5002 generic.go:334] "Generic (PLEG): container finished" podID="9281e6ad-d122-4c8d-aac8-256df9453697" containerID="51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76" exitCode=0 Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.901263 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerDied","Data":"51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76"} Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.905082 5002 generic.go:334] "Generic (PLEG): container finished" podID="763f3c28-d977-4ebc-81a1-e774138472c3" containerID="f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c" exitCode=0 Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.905122 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerDied","Data":"f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c"} Oct 14 09:09:29 crc kubenswrapper[5002]: I1014 09:09:29.905151 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerStarted","Data":"88dc6550a12932c59334207c9027ee76788b100378df4b83fdf05a8ed60d6d9d"} Oct 14 09:09:30 crc kubenswrapper[5002]: I1014 09:09:30.916634 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerStarted","Data":"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7"} Oct 14 09:09:31 crc kubenswrapper[5002]: E1014 09:09:31.701982 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9281e6ad_d122_4c8d_aac8_256df9453697.slice/crio-9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9281e6ad_d122_4c8d_aac8_256df9453697.slice/crio-conmon-9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:09:31 crc kubenswrapper[5002]: I1014 09:09:31.927078 5002 generic.go:334] "Generic (PLEG): container finished" podID="9281e6ad-d122-4c8d-aac8-256df9453697" containerID="9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef" exitCode=0 Oct 14 09:09:31 crc kubenswrapper[5002]: I1014 09:09:31.927149 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerDied","Data":"9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef"} Oct 14 09:09:32 crc kubenswrapper[5002]: I1014 09:09:32.938603 5002 generic.go:334] "Generic (PLEG): container finished" podID="763f3c28-d977-4ebc-81a1-e774138472c3" containerID="f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7" exitCode=0 Oct 14 09:09:32 crc kubenswrapper[5002]: I1014 09:09:32.938650 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerDied","Data":"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7"} Oct 14 09:09:33 crc kubenswrapper[5002]: I1014 09:09:33.948649 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerStarted","Data":"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d"} Oct 14 09:09:33 crc kubenswrapper[5002]: I1014 09:09:33.951003 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerStarted","Data":"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878"} Oct 14 09:09:33 crc kubenswrapper[5002]: I1014 09:09:33.977678 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gq2pn" podStartSLOduration=3.855508631 podStartE2EDuration="6.977655736s" podCreationTimestamp="2025-10-14 09:09:27 +0000 UTC" firstStartedPulling="2025-10-14 09:09:29.902627183 +0000 UTC m=+4702.883866625" lastFinishedPulling="2025-10-14 09:09:33.024774278 +0000 UTC m=+4706.006013730" observedRunningTime="2025-10-14 09:09:33.972295223 +0000 UTC m=+4706.953534675" watchObservedRunningTime="2025-10-14 09:09:33.977655736 +0000 UTC m=+4706.958895188" Oct 14 09:09:33 crc kubenswrapper[5002]: I1014 09:09:33.986820 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wk5x7" podStartSLOduration=2.242339102 podStartE2EDuration="5.98679988s" podCreationTimestamp="2025-10-14 09:09:28 +0000 UTC" firstStartedPulling="2025-10-14 09:09:29.908108689 +0000 UTC m=+4702.889348141" lastFinishedPulling="2025-10-14 09:09:33.652569467 +0000 UTC m=+4706.633808919" observedRunningTime="2025-10-14 09:09:33.986079122 +0000 UTC m=+4706.967318594" watchObservedRunningTime="2025-10-14 09:09:33.98679988 +0000 UTC m=+4706.968039332" Oct 14 09:09:37 crc kubenswrapper[5002]: I1014 09:09:37.532240 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:37 crc kubenswrapper[5002]: I1014 09:09:37.532894 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:37 crc kubenswrapper[5002]: I1014 09:09:37.601455 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:38 crc kubenswrapper[5002]: I1014 09:09:38.854274 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:38 crc kubenswrapper[5002]: I1014 09:09:38.854706 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:38 crc kubenswrapper[5002]: I1014 09:09:38.942405 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:39 crc kubenswrapper[5002]: I1014 09:09:39.082774 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:39 crc kubenswrapper[5002]: I1014 09:09:39.218515 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:09:39 crc kubenswrapper[5002]: I1014 09:09:39.219029 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:09:40 crc kubenswrapper[5002]: I1014 09:09:40.191162 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:41 crc kubenswrapper[5002]: I1014 09:09:41.009877 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wk5x7" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="registry-server" containerID="cri-o://a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878" gracePeriod=2 Oct 14 09:09:41 crc kubenswrapper[5002]: I1014 09:09:41.873976 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.011958 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzx4n\" (UniqueName: \"kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n\") pod \"763f3c28-d977-4ebc-81a1-e774138472c3\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.012366 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities\") pod \"763f3c28-d977-4ebc-81a1-e774138472c3\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.012395 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content\") pod \"763f3c28-d977-4ebc-81a1-e774138472c3\" (UID: \"763f3c28-d977-4ebc-81a1-e774138472c3\") " Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.014151 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities" (OuterVolumeSpecName: "utilities") pod "763f3c28-d977-4ebc-81a1-e774138472c3" (UID: "763f3c28-d977-4ebc-81a1-e774138472c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.019396 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n" (OuterVolumeSpecName: "kube-api-access-wzx4n") pod "763f3c28-d977-4ebc-81a1-e774138472c3" (UID: "763f3c28-d977-4ebc-81a1-e774138472c3"). InnerVolumeSpecName "kube-api-access-wzx4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.025174 5002 generic.go:334] "Generic (PLEG): container finished" podID="763f3c28-d977-4ebc-81a1-e774138472c3" containerID="a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878" exitCode=0 Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.025250 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerDied","Data":"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878"} Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.025364 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wk5x7" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.025381 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wk5x7" event={"ID":"763f3c28-d977-4ebc-81a1-e774138472c3","Type":"ContainerDied","Data":"88dc6550a12932c59334207c9027ee76788b100378df4b83fdf05a8ed60d6d9d"} Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.025400 5002 scope.go:117] "RemoveContainer" containerID="a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.057548 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "763f3c28-d977-4ebc-81a1-e774138472c3" (UID: "763f3c28-d977-4ebc-81a1-e774138472c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.074185 5002 scope.go:117] "RemoveContainer" containerID="f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.096640 5002 scope.go:117] "RemoveContainer" containerID="f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.114557 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzx4n\" (UniqueName: \"kubernetes.io/projected/763f3c28-d977-4ebc-81a1-e774138472c3-kube-api-access-wzx4n\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.114596 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.114605 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763f3c28-d977-4ebc-81a1-e774138472c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.138374 5002 scope.go:117] "RemoveContainer" containerID="a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878" Oct 14 09:09:42 crc kubenswrapper[5002]: E1014 09:09:42.139134 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878\": container with ID starting with a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878 not found: ID does not exist" containerID="a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.139188 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878"} err="failed to get container status \"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878\": rpc error: code = NotFound desc = could not find container \"a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878\": container with ID starting with a21e92647903c251ff184408ae37eb1cac3e4d02f064e6b56b94b14cb03ab878 not found: ID does not exist" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.139220 5002 scope.go:117] "RemoveContainer" containerID="f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7" Oct 14 09:09:42 crc kubenswrapper[5002]: E1014 09:09:42.139700 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7\": container with ID starting with f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7 not found: ID does not exist" containerID="f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.139738 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7"} err="failed to get container status \"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7\": rpc error: code = NotFound desc = could not find container \"f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7\": container with ID starting with f36ef7088daccbd5bca972de88550e1bcedd642912da42ef2fb6751863ff0cb7 not found: ID does not exist" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.139760 5002 scope.go:117] "RemoveContainer" containerID="f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c" Oct 14 09:09:42 crc kubenswrapper[5002]: E1014 09:09:42.140251 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c\": container with ID starting with f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c not found: ID does not exist" containerID="f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.140422 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c"} err="failed to get container status \"f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c\": rpc error: code = NotFound desc = could not find container \"f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c\": container with ID starting with f94d63da22df443112bc9111ae6b7f2901e9388650cd30d625ccdd3b0337a60c not found: ID does not exist" Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.372568 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:42 crc kubenswrapper[5002]: I1014 09:09:42.383732 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wk5x7"] Oct 14 09:09:43 crc kubenswrapper[5002]: I1014 09:09:43.734241 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" path="/var/lib/kubelet/pods/763f3c28-d977-4ebc-81a1-e774138472c3/volumes" Oct 14 09:09:47 crc kubenswrapper[5002]: I1014 09:09:47.599477 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:47 crc kubenswrapper[5002]: I1014 09:09:47.647561 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.084504 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gq2pn" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="registry-server" containerID="cri-o://81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d" gracePeriod=2 Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.847978 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.880601 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content\") pod \"9281e6ad-d122-4c8d-aac8-256df9453697\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.880790 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities\") pod \"9281e6ad-d122-4c8d-aac8-256df9453697\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.880987 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mbql\" (UniqueName: \"kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql\") pod \"9281e6ad-d122-4c8d-aac8-256df9453697\" (UID: \"9281e6ad-d122-4c8d-aac8-256df9453697\") " Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.881879 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities" (OuterVolumeSpecName: "utilities") pod "9281e6ad-d122-4c8d-aac8-256df9453697" (UID: "9281e6ad-d122-4c8d-aac8-256df9453697"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.887013 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql" (OuterVolumeSpecName: "kube-api-access-4mbql") pod "9281e6ad-d122-4c8d-aac8-256df9453697" (UID: "9281e6ad-d122-4c8d-aac8-256df9453697"). InnerVolumeSpecName "kube-api-access-4mbql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.911088 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9281e6ad-d122-4c8d-aac8-256df9453697" (UID: "9281e6ad-d122-4c8d-aac8-256df9453697"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.982927 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.982964 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mbql\" (UniqueName: \"kubernetes.io/projected/9281e6ad-d122-4c8d-aac8-256df9453697-kube-api-access-4mbql\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:48 crc kubenswrapper[5002]: I1014 09:09:48.982978 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9281e6ad-d122-4c8d-aac8-256df9453697-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.093892 5002 generic.go:334] "Generic (PLEG): container finished" podID="9281e6ad-d122-4c8d-aac8-256df9453697" containerID="81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d" exitCode=0 Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.094027 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gq2pn" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.094230 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerDied","Data":"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d"} Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.094692 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gq2pn" event={"ID":"9281e6ad-d122-4c8d-aac8-256df9453697","Type":"ContainerDied","Data":"c7ed5204751e7956d0b6829defecff6cbbec0ec13935381482162ca373d813f9"} Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.094782 5002 scope.go:117] "RemoveContainer" containerID="81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.114962 5002 scope.go:117] "RemoveContainer" containerID="9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.144047 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.148075 5002 scope.go:117] "RemoveContainer" containerID="51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.155221 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gq2pn"] Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.185082 5002 scope.go:117] "RemoveContainer" containerID="81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d" Oct 14 09:09:49 crc kubenswrapper[5002]: E1014 09:09:49.190899 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d\": container with ID starting with 81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d not found: ID does not exist" containerID="81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.191045 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d"} err="failed to get container status \"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d\": rpc error: code = NotFound desc = could not find container \"81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d\": container with ID starting with 81bb7f14a59a743850eb417102413dfba2bd8aaecc3b55ce956e4d868446064d not found: ID does not exist" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.191187 5002 scope.go:117] "RemoveContainer" containerID="9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef" Oct 14 09:09:49 crc kubenswrapper[5002]: E1014 09:09:49.191924 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef\": container with ID starting with 9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef not found: ID does not exist" containerID="9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.191995 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef"} err="failed to get container status \"9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef\": rpc error: code = NotFound desc = could not find container \"9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef\": container with ID starting with 9c4d76f3131d74734c5bb1307aed4002de3714561abcc6533ff7a3ce7e5000ef not found: ID does not exist" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.192038 5002 scope.go:117] "RemoveContainer" containerID="51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76" Oct 14 09:09:49 crc kubenswrapper[5002]: E1014 09:09:49.192572 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76\": container with ID starting with 51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76 not found: ID does not exist" containerID="51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.192623 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76"} err="failed to get container status \"51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76\": rpc error: code = NotFound desc = could not find container \"51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76\": container with ID starting with 51983cabd7b8a8380886eae04e7b68aded2b8df8a960aa49c038baae9db28b76 not found: ID does not exist" Oct 14 09:09:49 crc kubenswrapper[5002]: I1014 09:09:49.736139 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" path="/var/lib/kubelet/pods/9281e6ad-d122-4c8d-aac8-256df9453697/volumes" Oct 14 09:10:09 crc kubenswrapper[5002]: I1014 09:10:09.218169 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:10:09 crc kubenswrapper[5002]: I1014 09:10:09.218829 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:10:09 crc kubenswrapper[5002]: I1014 09:10:09.218903 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:10:09 crc kubenswrapper[5002]: I1014 09:10:09.219795 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:10:09 crc kubenswrapper[5002]: I1014 09:10:09.219888 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b" gracePeriod=600 Oct 14 09:10:10 crc kubenswrapper[5002]: I1014 09:10:10.322919 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b" exitCode=0 Oct 14 09:10:10 crc kubenswrapper[5002]: I1014 09:10:10.323010 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b"} Oct 14 09:10:10 crc kubenswrapper[5002]: I1014 09:10:10.323619 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5"} Oct 14 09:10:10 crc kubenswrapper[5002]: I1014 09:10:10.323649 5002 scope.go:117] "RemoveContainer" containerID="0090a048b7e1df7e130cebf78f7c32b476b6c7732881ed6cef2ff974a21573ca" Oct 14 09:12:09 crc kubenswrapper[5002]: I1014 09:12:09.219655 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:12:09 crc kubenswrapper[5002]: I1014 09:12:09.220313 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:12:39 crc kubenswrapper[5002]: I1014 09:12:39.218073 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:12:39 crc kubenswrapper[5002]: I1014 09:12:39.218673 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:13:09 crc kubenswrapper[5002]: I1014 09:13:09.218550 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:13:09 crc kubenswrapper[5002]: I1014 09:13:09.219177 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:13:09 crc kubenswrapper[5002]: I1014 09:13:09.219238 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:13:09 crc kubenswrapper[5002]: I1014 09:13:09.220032 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:13:09 crc kubenswrapper[5002]: I1014 09:13:09.220093 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" gracePeriod=600 Oct 14 09:13:09 crc kubenswrapper[5002]: E1014 09:13:09.355707 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:13:10 crc kubenswrapper[5002]: I1014 09:13:10.042559 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" exitCode=0 Oct 14 09:13:10 crc kubenswrapper[5002]: I1014 09:13:10.042620 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5"} Oct 14 09:13:10 crc kubenswrapper[5002]: I1014 09:13:10.042713 5002 scope.go:117] "RemoveContainer" containerID="cf71de9734d6adc01f3f234434b08690fe1c0bd82f71d9bc6cb22faf5154295b" Oct 14 09:13:10 crc kubenswrapper[5002]: I1014 09:13:10.043799 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:13:10 crc kubenswrapper[5002]: E1014 09:13:10.047889 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:13:23 crc kubenswrapper[5002]: I1014 09:13:23.720935 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:13:23 crc kubenswrapper[5002]: E1014 09:13:23.721760 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:13:36 crc kubenswrapper[5002]: I1014 09:13:36.720267 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:13:36 crc kubenswrapper[5002]: E1014 09:13:36.721020 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:13:49 crc kubenswrapper[5002]: I1014 09:13:49.721021 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:13:49 crc kubenswrapper[5002]: E1014 09:13:49.721878 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:14:00 crc kubenswrapper[5002]: I1014 09:14:00.723104 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:14:00 crc kubenswrapper[5002]: E1014 09:14:00.724210 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:14:14 crc kubenswrapper[5002]: I1014 09:14:14.720331 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:14:14 crc kubenswrapper[5002]: E1014 09:14:14.721044 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:14:27 crc kubenswrapper[5002]: I1014 09:14:27.726578 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:14:27 crc kubenswrapper[5002]: E1014 09:14:27.727559 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:14:39 crc kubenswrapper[5002]: I1014 09:14:39.721696 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:14:39 crc kubenswrapper[5002]: E1014 09:14:39.722668 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:14:51 crc kubenswrapper[5002]: I1014 09:14:51.721300 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:14:51 crc kubenswrapper[5002]: E1014 09:14:51.722062 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.145517 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg"] Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146475 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146492 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146507 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146514 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146538 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146547 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146561 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146568 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146586 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146593 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[5002]: E1014 09:15:00.146607 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.146615 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.152683 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="763f3c28-d977-4ebc-81a1-e774138472c3" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.152791 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9281e6ad-d122-4c8d-aac8-256df9453697" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.154589 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.183274 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.183493 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.200587 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg"] Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.301015 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpjr6\" (UniqueName: \"kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.301122 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.301211 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.402402 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.402476 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpjr6\" (UniqueName: \"kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.402643 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.404560 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.456350 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.457120 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpjr6\" (UniqueName: \"kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6\") pod \"collect-profiles-29340555-hpxzg\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:00 crc kubenswrapper[5002]: I1014 09:15:00.518552 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:01 crc kubenswrapper[5002]: I1014 09:15:01.006469 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg"] Oct 14 09:15:01 crc kubenswrapper[5002]: I1014 09:15:01.110290 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" event={"ID":"fc7a8845-3a0c-4f98-9605-6547d7a52746","Type":"ContainerStarted","Data":"ba1b2a83a771546c76d04f72967b8853e7643b8ec79f3ce841dba1603921c451"} Oct 14 09:15:02 crc kubenswrapper[5002]: I1014 09:15:02.121959 5002 generic.go:334] "Generic (PLEG): container finished" podID="fc7a8845-3a0c-4f98-9605-6547d7a52746" containerID="cb41609ad3f7fe3f33f33ef5f7e6b9122b9394414eddb5a7b69ceb9d88250bcd" exitCode=0 Oct 14 09:15:02 crc kubenswrapper[5002]: I1014 09:15:02.122074 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" event={"ID":"fc7a8845-3a0c-4f98-9605-6547d7a52746","Type":"ContainerDied","Data":"cb41609ad3f7fe3f33f33ef5f7e6b9122b9394414eddb5a7b69ceb9d88250bcd"} Oct 14 09:15:02 crc kubenswrapper[5002]: I1014 09:15:02.721005 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:15:02 crc kubenswrapper[5002]: E1014 09:15:02.721682 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.657508 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.772809 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume\") pod \"fc7a8845-3a0c-4f98-9605-6547d7a52746\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.772900 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume\") pod \"fc7a8845-3a0c-4f98-9605-6547d7a52746\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.773148 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpjr6\" (UniqueName: \"kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6\") pod \"fc7a8845-3a0c-4f98-9605-6547d7a52746\" (UID: \"fc7a8845-3a0c-4f98-9605-6547d7a52746\") " Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.773599 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume" (OuterVolumeSpecName: "config-volume") pod "fc7a8845-3a0c-4f98-9605-6547d7a52746" (UID: "fc7a8845-3a0c-4f98-9605-6547d7a52746"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.779466 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6" (OuterVolumeSpecName: "kube-api-access-zpjr6") pod "fc7a8845-3a0c-4f98-9605-6547d7a52746" (UID: "fc7a8845-3a0c-4f98-9605-6547d7a52746"). InnerVolumeSpecName "kube-api-access-zpjr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.792235 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fc7a8845-3a0c-4f98-9605-6547d7a52746" (UID: "fc7a8845-3a0c-4f98-9605-6547d7a52746"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.876087 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpjr6\" (UniqueName: \"kubernetes.io/projected/fc7a8845-3a0c-4f98-9605-6547d7a52746-kube-api-access-zpjr6\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.876116 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc7a8845-3a0c-4f98-9605-6547d7a52746-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:03 crc kubenswrapper[5002]: I1014 09:15:03.876127 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc7a8845-3a0c-4f98-9605-6547d7a52746-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:04 crc kubenswrapper[5002]: I1014 09:15:04.144177 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" event={"ID":"fc7a8845-3a0c-4f98-9605-6547d7a52746","Type":"ContainerDied","Data":"ba1b2a83a771546c76d04f72967b8853e7643b8ec79f3ce841dba1603921c451"} Oct 14 09:15:04 crc kubenswrapper[5002]: I1014 09:15:04.144547 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba1b2a83a771546c76d04f72967b8853e7643b8ec79f3ce841dba1603921c451" Oct 14 09:15:04 crc kubenswrapper[5002]: I1014 09:15:04.144326 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg" Oct 14 09:15:04 crc kubenswrapper[5002]: I1014 09:15:04.740675 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss"] Oct 14 09:15:04 crc kubenswrapper[5002]: I1014 09:15:04.749117 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-xrbss"] Oct 14 09:15:05 crc kubenswrapper[5002]: I1014 09:15:05.730077 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b52361-3089-4ec1-a9aa-c29bc75a7007" path="/var/lib/kubelet/pods/82b52361-3089-4ec1-a9aa-c29bc75a7007/volumes" Oct 14 09:15:17 crc kubenswrapper[5002]: I1014 09:15:17.730991 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:15:17 crc kubenswrapper[5002]: E1014 09:15:17.731813 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:15:28 crc kubenswrapper[5002]: I1014 09:15:28.721145 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:15:28 crc kubenswrapper[5002]: E1014 09:15:28.722334 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:15:29 crc kubenswrapper[5002]: I1014 09:15:29.789366 5002 scope.go:117] "RemoveContainer" containerID="f1ba0fa315f025033787adb1b32273076d78ae390483a9c0a091c87e253a91ea" Oct 14 09:15:42 crc kubenswrapper[5002]: I1014 09:15:42.721353 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:15:42 crc kubenswrapper[5002]: E1014 09:15:42.722145 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:15:53 crc kubenswrapper[5002]: I1014 09:15:53.721392 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:15:53 crc kubenswrapper[5002]: E1014 09:15:53.722710 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:16:05 crc kubenswrapper[5002]: I1014 09:16:05.720602 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:16:05 crc kubenswrapper[5002]: E1014 09:16:05.721452 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:16:18 crc kubenswrapper[5002]: I1014 09:16:18.720992 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:16:18 crc kubenswrapper[5002]: E1014 09:16:18.722182 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:16:33 crc kubenswrapper[5002]: I1014 09:16:33.721833 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:16:33 crc kubenswrapper[5002]: E1014 09:16:33.722683 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:16:46 crc kubenswrapper[5002]: I1014 09:16:46.721018 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:16:46 crc kubenswrapper[5002]: E1014 09:16:46.721729 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:16:57 crc kubenswrapper[5002]: I1014 09:16:57.725865 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:16:57 crc kubenswrapper[5002]: E1014 09:16:57.726640 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:17:11 crc kubenswrapper[5002]: I1014 09:17:11.720467 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:17:11 crc kubenswrapper[5002]: E1014 09:17:11.721436 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:17:24 crc kubenswrapper[5002]: I1014 09:17:24.720935 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:17:24 crc kubenswrapper[5002]: E1014 09:17:24.721680 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:17:36 crc kubenswrapper[5002]: I1014 09:17:36.720388 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:17:36 crc kubenswrapper[5002]: E1014 09:17:36.721118 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:17:51 crc kubenswrapper[5002]: I1014 09:17:51.721269 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:17:51 crc kubenswrapper[5002]: E1014 09:17:51.722562 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:18:05 crc kubenswrapper[5002]: I1014 09:18:05.720720 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:18:05 crc kubenswrapper[5002]: E1014 09:18:05.721631 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:18:18 crc kubenswrapper[5002]: I1014 09:18:18.721046 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:18:20 crc kubenswrapper[5002]: I1014 09:18:20.134254 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185"} Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.395079 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:30 crc kubenswrapper[5002]: E1014 09:18:30.396114 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc7a8845-3a0c-4f98-9605-6547d7a52746" containerName="collect-profiles" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.396133 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc7a8845-3a0c-4f98-9605-6547d7a52746" containerName="collect-profiles" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.396426 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc7a8845-3a0c-4f98-9605-6547d7a52746" containerName="collect-profiles" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.398103 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.405597 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.482496 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.482598 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9mz6\" (UniqueName: \"kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.482760 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.584902 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.584978 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.585033 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9mz6\" (UniqueName: \"kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.585514 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.585523 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.606276 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9mz6\" (UniqueName: \"kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6\") pod \"redhat-operators-p95rk\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:30 crc kubenswrapper[5002]: I1014 09:18:30.729751 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:31 crc kubenswrapper[5002]: I1014 09:18:31.217558 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:32 crc kubenswrapper[5002]: I1014 09:18:32.256832 5002 generic.go:334] "Generic (PLEG): container finished" podID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerID="429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835" exitCode=0 Oct 14 09:18:32 crc kubenswrapper[5002]: I1014 09:18:32.257401 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerDied","Data":"429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835"} Oct 14 09:18:32 crc kubenswrapper[5002]: I1014 09:18:32.257455 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerStarted","Data":"78665ac302aadbc9c74bb98c1f365c311df2bbc385db4ccd5449a57837d87a66"} Oct 14 09:18:32 crc kubenswrapper[5002]: I1014 09:18:32.260047 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:18:33 crc kubenswrapper[5002]: I1014 09:18:33.272307 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerStarted","Data":"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7"} Oct 14 09:18:37 crc kubenswrapper[5002]: I1014 09:18:37.308401 5002 generic.go:334] "Generic (PLEG): container finished" podID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerID="e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7" exitCode=0 Oct 14 09:18:37 crc kubenswrapper[5002]: I1014 09:18:37.308433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerDied","Data":"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7"} Oct 14 09:18:38 crc kubenswrapper[5002]: I1014 09:18:38.324468 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerStarted","Data":"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf"} Oct 14 09:18:38 crc kubenswrapper[5002]: I1014 09:18:38.343519 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p95rk" podStartSLOduration=2.614984666 podStartE2EDuration="8.343497214s" podCreationTimestamp="2025-10-14 09:18:30 +0000 UTC" firstStartedPulling="2025-10-14 09:18:32.259758897 +0000 UTC m=+5245.240998349" lastFinishedPulling="2025-10-14 09:18:37.988271445 +0000 UTC m=+5250.969510897" observedRunningTime="2025-10-14 09:18:38.342150858 +0000 UTC m=+5251.323390340" watchObservedRunningTime="2025-10-14 09:18:38.343497214 +0000 UTC m=+5251.324736666" Oct 14 09:18:40 crc kubenswrapper[5002]: I1014 09:18:40.730407 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:40 crc kubenswrapper[5002]: I1014 09:18:40.730707 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:41 crc kubenswrapper[5002]: I1014 09:18:41.999732 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p95rk" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="registry-server" probeResult="failure" output=< Oct 14 09:18:41 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:18:41 crc kubenswrapper[5002]: > Oct 14 09:18:50 crc kubenswrapper[5002]: I1014 09:18:50.822622 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:50 crc kubenswrapper[5002]: I1014 09:18:50.889910 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:51 crc kubenswrapper[5002]: I1014 09:18:51.074906 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:52 crc kubenswrapper[5002]: I1014 09:18:52.453053 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p95rk" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="registry-server" containerID="cri-o://f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf" gracePeriod=2 Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.089506 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.145442 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9mz6\" (UniqueName: \"kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6\") pod \"3e7b4a23-c805-4649-8345-82a03d61cddb\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.145522 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content\") pod \"3e7b4a23-c805-4649-8345-82a03d61cddb\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.145653 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities\") pod \"3e7b4a23-c805-4649-8345-82a03d61cddb\" (UID: \"3e7b4a23-c805-4649-8345-82a03d61cddb\") " Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.151626 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities" (OuterVolumeSpecName: "utilities") pod "3e7b4a23-c805-4649-8345-82a03d61cddb" (UID: "3e7b4a23-c805-4649-8345-82a03d61cddb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.167138 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6" (OuterVolumeSpecName: "kube-api-access-s9mz6") pod "3e7b4a23-c805-4649-8345-82a03d61cddb" (UID: "3e7b4a23-c805-4649-8345-82a03d61cddb"). InnerVolumeSpecName "kube-api-access-s9mz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.247489 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.247524 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9mz6\" (UniqueName: \"kubernetes.io/projected/3e7b4a23-c805-4649-8345-82a03d61cddb-kube-api-access-s9mz6\") on node \"crc\" DevicePath \"\"" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.249192 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e7b4a23-c805-4649-8345-82a03d61cddb" (UID: "3e7b4a23-c805-4649-8345-82a03d61cddb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.349599 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e7b4a23-c805-4649-8345-82a03d61cddb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.463706 5002 generic.go:334] "Generic (PLEG): container finished" podID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerID="f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf" exitCode=0 Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.463752 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerDied","Data":"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf"} Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.463782 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p95rk" event={"ID":"3e7b4a23-c805-4649-8345-82a03d61cddb","Type":"ContainerDied","Data":"78665ac302aadbc9c74bb98c1f365c311df2bbc385db4ccd5449a57837d87a66"} Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.463802 5002 scope.go:117] "RemoveContainer" containerID="f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.463967 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p95rk" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.503545 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.505476 5002 scope.go:117] "RemoveContainer" containerID="e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.513482 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p95rk"] Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.525870 5002 scope.go:117] "RemoveContainer" containerID="429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.576370 5002 scope.go:117] "RemoveContainer" containerID="f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf" Oct 14 09:18:53 crc kubenswrapper[5002]: E1014 09:18:53.577096 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf\": container with ID starting with f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf not found: ID does not exist" containerID="f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.577152 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf"} err="failed to get container status \"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf\": rpc error: code = NotFound desc = could not find container \"f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf\": container with ID starting with f8bb4ed54cfd3c23043275b7e36ab23dbf82d20104408b9020fa278e9893b2bf not found: ID does not exist" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.577182 5002 scope.go:117] "RemoveContainer" containerID="e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7" Oct 14 09:18:53 crc kubenswrapper[5002]: E1014 09:18:53.577547 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7\": container with ID starting with e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7 not found: ID does not exist" containerID="e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.577773 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7"} err="failed to get container status \"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7\": rpc error: code = NotFound desc = could not find container \"e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7\": container with ID starting with e84e20568e2d82251e2ec18c1eef1a4bbc2075c196076bbf93117e45a48490d7 not found: ID does not exist" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.577882 5002 scope.go:117] "RemoveContainer" containerID="429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835" Oct 14 09:18:53 crc kubenswrapper[5002]: E1014 09:18:53.578324 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835\": container with ID starting with 429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835 not found: ID does not exist" containerID="429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.578412 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835"} err="failed to get container status \"429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835\": rpc error: code = NotFound desc = could not find container \"429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835\": container with ID starting with 429c858cfd04ebcfb9c497b988b5a2eff50d069c3848eba899b454fa82f62835 not found: ID does not exist" Oct 14 09:18:53 crc kubenswrapper[5002]: I1014 09:18:53.733538 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" path="/var/lib/kubelet/pods/3e7b4a23-c805-4649-8345-82a03d61cddb/volumes" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.953958 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:23 crc kubenswrapper[5002]: E1014 09:19:23.955337 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="extract-utilities" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.955362 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="extract-utilities" Oct 14 09:19:23 crc kubenswrapper[5002]: E1014 09:19:23.955415 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="extract-content" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.955429 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="extract-content" Oct 14 09:19:23 crc kubenswrapper[5002]: E1014 09:19:23.955456 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="registry-server" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.955469 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="registry-server" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.955875 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e7b4a23-c805-4649-8345-82a03d61cddb" containerName="registry-server" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.958245 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:23 crc kubenswrapper[5002]: I1014 09:19:23.962384 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.024602 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.024714 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.024783 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2j8n\" (UniqueName: \"kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.126637 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.126950 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.126971 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.127156 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2j8n\" (UniqueName: \"kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.127531 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.163377 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2j8n\" (UniqueName: \"kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n\") pod \"community-operators-nwbj6\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.281889 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:24 crc kubenswrapper[5002]: I1014 09:19:24.808084 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:25 crc kubenswrapper[5002]: I1014 09:19:25.791383 5002 generic.go:334] "Generic (PLEG): container finished" podID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerID="6d8f3a2f95d307618987e21272048c37e1f613d813f6cb74b4a3c0bde8aa3699" exitCode=0 Oct 14 09:19:25 crc kubenswrapper[5002]: I1014 09:19:25.791679 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerDied","Data":"6d8f3a2f95d307618987e21272048c37e1f613d813f6cb74b4a3c0bde8aa3699"} Oct 14 09:19:25 crc kubenswrapper[5002]: I1014 09:19:25.791711 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerStarted","Data":"cd158d1287e8ba39f592fab76cac26e82f7af901cc12977917cd6734191427ac"} Oct 14 09:19:27 crc kubenswrapper[5002]: I1014 09:19:27.818890 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerStarted","Data":"ad5ab708a29970468e1e8ee5693550ee7a1eb9fefa72d53a288a2585813ea525"} Oct 14 09:19:28 crc kubenswrapper[5002]: I1014 09:19:28.839337 5002 generic.go:334] "Generic (PLEG): container finished" podID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerID="ad5ab708a29970468e1e8ee5693550ee7a1eb9fefa72d53a288a2585813ea525" exitCode=0 Oct 14 09:19:28 crc kubenswrapper[5002]: I1014 09:19:28.839494 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerDied","Data":"ad5ab708a29970468e1e8ee5693550ee7a1eb9fefa72d53a288a2585813ea525"} Oct 14 09:19:30 crc kubenswrapper[5002]: I1014 09:19:30.860199 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerStarted","Data":"1bf8c30dc973e9a0ce44bf7cb4e02d450a1ead952631553b58dbfbd6b3ee6a78"} Oct 14 09:19:30 crc kubenswrapper[5002]: I1014 09:19:30.887127 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwbj6" podStartSLOduration=4.132015761 podStartE2EDuration="7.887108743s" podCreationTimestamp="2025-10-14 09:19:23 +0000 UTC" firstStartedPulling="2025-10-14 09:19:25.803478442 +0000 UTC m=+5298.784717904" lastFinishedPulling="2025-10-14 09:19:29.558571434 +0000 UTC m=+5302.539810886" observedRunningTime="2025-10-14 09:19:30.880232789 +0000 UTC m=+5303.861472251" watchObservedRunningTime="2025-10-14 09:19:30.887108743 +0000 UTC m=+5303.868348195" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.584710 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.588297 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.616140 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.714381 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.714815 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.714886 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2swq\" (UniqueName: \"kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.816440 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.816553 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.816579 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2swq\" (UniqueName: \"kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.817195 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.817226 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.839262 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2swq\" (UniqueName: \"kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq\") pod \"redhat-marketplace-pdgg7\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:32 crc kubenswrapper[5002]: I1014 09:19:32.916529 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:33 crc kubenswrapper[5002]: I1014 09:19:33.372245 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:33 crc kubenswrapper[5002]: W1014 09:19:33.387040 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b509d23_ef36_4385_ace3_4f13561ff06e.slice/crio-964ea61a452bf16084efd3ab99b3f9ed6cced815c7fae360fca170a7c1d5236c WatchSource:0}: Error finding container 964ea61a452bf16084efd3ab99b3f9ed6cced815c7fae360fca170a7c1d5236c: Status 404 returned error can't find the container with id 964ea61a452bf16084efd3ab99b3f9ed6cced815c7fae360fca170a7c1d5236c Oct 14 09:19:33 crc kubenswrapper[5002]: I1014 09:19:33.892304 5002 generic.go:334] "Generic (PLEG): container finished" podID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerID="28403219bc6bd4a20eee515ea77b460eea579537876c5d5fb3b2e695673bdbe7" exitCode=0 Oct 14 09:19:33 crc kubenswrapper[5002]: I1014 09:19:33.892382 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerDied","Data":"28403219bc6bd4a20eee515ea77b460eea579537876c5d5fb3b2e695673bdbe7"} Oct 14 09:19:33 crc kubenswrapper[5002]: I1014 09:19:33.892640 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerStarted","Data":"964ea61a452bf16084efd3ab99b3f9ed6cced815c7fae360fca170a7c1d5236c"} Oct 14 09:19:34 crc kubenswrapper[5002]: I1014 09:19:34.283265 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:34 crc kubenswrapper[5002]: I1014 09:19:34.283696 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:34 crc kubenswrapper[5002]: I1014 09:19:34.347308 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:34 crc kubenswrapper[5002]: I1014 09:19:34.903532 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerStarted","Data":"fc3eaebff89093da20c6b623e32584a5d04a0227e8e8851a5ec977b59fa771a6"} Oct 14 09:19:35 crc kubenswrapper[5002]: I1014 09:19:35.916372 5002 generic.go:334] "Generic (PLEG): container finished" podID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerID="fc3eaebff89093da20c6b623e32584a5d04a0227e8e8851a5ec977b59fa771a6" exitCode=0 Oct 14 09:19:35 crc kubenswrapper[5002]: I1014 09:19:35.916444 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerDied","Data":"fc3eaebff89093da20c6b623e32584a5d04a0227e8e8851a5ec977b59fa771a6"} Oct 14 09:19:36 crc kubenswrapper[5002]: I1014 09:19:36.936400 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerStarted","Data":"b8e9c8a882489af973f577b6ba37ddff32526ab82c4e2ce5869a0c82345c7e14"} Oct 14 09:19:36 crc kubenswrapper[5002]: I1014 09:19:36.960289 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pdgg7" podStartSLOduration=2.267931666 podStartE2EDuration="4.960248246s" podCreationTimestamp="2025-10-14 09:19:32 +0000 UTC" firstStartedPulling="2025-10-14 09:19:33.895372823 +0000 UTC m=+5306.876612305" lastFinishedPulling="2025-10-14 09:19:36.587689433 +0000 UTC m=+5309.568928885" observedRunningTime="2025-10-14 09:19:36.960073022 +0000 UTC m=+5309.941312484" watchObservedRunningTime="2025-10-14 09:19:36.960248246 +0000 UTC m=+5309.941487698" Oct 14 09:19:42 crc kubenswrapper[5002]: I1014 09:19:42.918696 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:42 crc kubenswrapper[5002]: I1014 09:19:42.919322 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:42 crc kubenswrapper[5002]: I1014 09:19:42.979062 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:43 crc kubenswrapper[5002]: I1014 09:19:43.060121 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:43 crc kubenswrapper[5002]: I1014 09:19:43.218572 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:44 crc kubenswrapper[5002]: I1014 09:19:44.362707 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:45 crc kubenswrapper[5002]: I1014 09:19:45.016473 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pdgg7" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="registry-server" containerID="cri-o://b8e9c8a882489af973f577b6ba37ddff32526ab82c4e2ce5869a0c82345c7e14" gracePeriod=2 Oct 14 09:19:45 crc kubenswrapper[5002]: I1014 09:19:45.620980 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:45 crc kubenswrapper[5002]: I1014 09:19:45.621510 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwbj6" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="registry-server" containerID="cri-o://1bf8c30dc973e9a0ce44bf7cb4e02d450a1ead952631553b58dbfbd6b3ee6a78" gracePeriod=2 Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.031936 5002 generic.go:334] "Generic (PLEG): container finished" podID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerID="1bf8c30dc973e9a0ce44bf7cb4e02d450a1ead952631553b58dbfbd6b3ee6a78" exitCode=0 Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.032106 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerDied","Data":"1bf8c30dc973e9a0ce44bf7cb4e02d450a1ead952631553b58dbfbd6b3ee6a78"} Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.037490 5002 generic.go:334] "Generic (PLEG): container finished" podID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerID="b8e9c8a882489af973f577b6ba37ddff32526ab82c4e2ce5869a0c82345c7e14" exitCode=0 Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.037530 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerDied","Data":"b8e9c8a882489af973f577b6ba37ddff32526ab82c4e2ce5869a0c82345c7e14"} Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.344139 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.351052 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.499808 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2swq\" (UniqueName: \"kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq\") pod \"9b509d23-ef36-4385-ace3-4f13561ff06e\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.499910 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2j8n\" (UniqueName: \"kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n\") pod \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.500010 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content\") pod \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.500056 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities\") pod \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\" (UID: \"401c82bc-1ee6-4fc4-a143-27b7e733b17f\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.500169 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities\") pod \"9b509d23-ef36-4385-ace3-4f13561ff06e\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.500375 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content\") pod \"9b509d23-ef36-4385-ace3-4f13561ff06e\" (UID: \"9b509d23-ef36-4385-ace3-4f13561ff06e\") " Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.500930 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities" (OuterVolumeSpecName: "utilities") pod "401c82bc-1ee6-4fc4-a143-27b7e733b17f" (UID: "401c82bc-1ee6-4fc4-a143-27b7e733b17f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.501289 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities" (OuterVolumeSpecName: "utilities") pod "9b509d23-ef36-4385-ace3-4f13561ff06e" (UID: "9b509d23-ef36-4385-ace3-4f13561ff06e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.505686 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.505947 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.523169 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b509d23-ef36-4385-ace3-4f13561ff06e" (UID: "9b509d23-ef36-4385-ace3-4f13561ff06e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.547322 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "401c82bc-1ee6-4fc4-a143-27b7e733b17f" (UID: "401c82bc-1ee6-4fc4-a143-27b7e733b17f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.609406 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/401c82bc-1ee6-4fc4-a143-27b7e733b17f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:46 crc kubenswrapper[5002]: I1014 09:19:46.609448 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b509d23-ef36-4385-ace3-4f13561ff06e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.052755 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pdgg7" event={"ID":"9b509d23-ef36-4385-ace3-4f13561ff06e","Type":"ContainerDied","Data":"964ea61a452bf16084efd3ab99b3f9ed6cced815c7fae360fca170a7c1d5236c"} Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.052818 5002 scope.go:117] "RemoveContainer" containerID="b8e9c8a882489af973f577b6ba37ddff32526ab82c4e2ce5869a0c82345c7e14" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.052984 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pdgg7" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.058572 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwbj6" event={"ID":"401c82bc-1ee6-4fc4-a143-27b7e733b17f","Type":"ContainerDied","Data":"cd158d1287e8ba39f592fab76cac26e82f7af901cc12977917cd6734191427ac"} Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.058619 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwbj6" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.082084 5002 scope.go:117] "RemoveContainer" containerID="fc3eaebff89093da20c6b623e32584a5d04a0227e8e8851a5ec977b59fa771a6" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.155672 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n" (OuterVolumeSpecName: "kube-api-access-j2j8n") pod "401c82bc-1ee6-4fc4-a143-27b7e733b17f" (UID: "401c82bc-1ee6-4fc4-a143-27b7e733b17f"). InnerVolumeSpecName "kube-api-access-j2j8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.222326 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2j8n\" (UniqueName: \"kubernetes.io/projected/401c82bc-1ee6-4fc4-a143-27b7e733b17f-kube-api-access-j2j8n\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.253098 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq" (OuterVolumeSpecName: "kube-api-access-f2swq") pod "9b509d23-ef36-4385-ace3-4f13561ff06e" (UID: "9b509d23-ef36-4385-ace3-4f13561ff06e"). InnerVolumeSpecName "kube-api-access-f2swq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.280564 5002 scope.go:117] "RemoveContainer" containerID="28403219bc6bd4a20eee515ea77b460eea579537876c5d5fb3b2e695673bdbe7" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.324725 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2swq\" (UniqueName: \"kubernetes.io/projected/9b509d23-ef36-4385-ace3-4f13561ff06e-kube-api-access-f2swq\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.329615 5002 scope.go:117] "RemoveContainer" containerID="1bf8c30dc973e9a0ce44bf7cb4e02d450a1ead952631553b58dbfbd6b3ee6a78" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.359875 5002 scope.go:117] "RemoveContainer" containerID="ad5ab708a29970468e1e8ee5693550ee7a1eb9fefa72d53a288a2585813ea525" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.403277 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.408367 5002 scope.go:117] "RemoveContainer" containerID="6d8f3a2f95d307618987e21272048c37e1f613d813f6cb74b4a3c0bde8aa3699" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.413948 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pdgg7"] Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.436529 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.452792 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwbj6"] Oct 14 09:19:47 crc kubenswrapper[5002]: E1014 09:19:47.613559 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod401c82bc_1ee6_4fc4_a143_27b7e733b17f.slice/crio-cd158d1287e8ba39f592fab76cac26e82f7af901cc12977917cd6734191427ac\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod401c82bc_1ee6_4fc4_a143_27b7e733b17f.slice\": RecentStats: unable to find data in memory cache]" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.734639 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" path="/var/lib/kubelet/pods/401c82bc-1ee6-4fc4-a143-27b7e733b17f/volumes" Oct 14 09:19:47 crc kubenswrapper[5002]: I1014 09:19:47.735624 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" path="/var/lib/kubelet/pods/9b509d23-ef36-4385-ace3-4f13561ff06e/volumes" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.745238 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746461 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="extract-utilities" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746484 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="extract-utilities" Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746520 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="extract-content" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746532 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="extract-content" Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746554 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="extract-utilities" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746567 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="extract-utilities" Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746592 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746606 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746628 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746640 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: E1014 09:20:15.746675 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="extract-content" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.746689 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="extract-content" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.747022 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="401c82bc-1ee6-4fc4-a143-27b7e733b17f" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.747065 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b509d23-ef36-4385-ace3-4f13561ff06e" containerName="registry-server" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.749354 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.763291 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.823635 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.823726 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf9hn\" (UniqueName: \"kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.823872 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.925258 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.925316 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf9hn\" (UniqueName: \"kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.925392 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.925818 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.925867 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:15 crc kubenswrapper[5002]: I1014 09:20:15.947452 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf9hn\" (UniqueName: \"kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn\") pod \"certified-operators-g86rw\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:16 crc kubenswrapper[5002]: I1014 09:20:16.073279 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:16 crc kubenswrapper[5002]: I1014 09:20:16.583983 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:17 crc kubenswrapper[5002]: I1014 09:20:17.387928 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerID="938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034" exitCode=0 Oct 14 09:20:17 crc kubenswrapper[5002]: I1014 09:20:17.388180 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerDied","Data":"938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034"} Oct 14 09:20:17 crc kubenswrapper[5002]: I1014 09:20:17.388211 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerStarted","Data":"b05da18433599035bf6cec4c6c2958664315baf9fb765b3423f0c348b2d446a8"} Oct 14 09:20:19 crc kubenswrapper[5002]: I1014 09:20:19.411269 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerStarted","Data":"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541"} Oct 14 09:20:20 crc kubenswrapper[5002]: I1014 09:20:20.424359 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerID="17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541" exitCode=0 Oct 14 09:20:20 crc kubenswrapper[5002]: I1014 09:20:20.424433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerDied","Data":"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541"} Oct 14 09:20:21 crc kubenswrapper[5002]: I1014 09:20:21.435798 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerStarted","Data":"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf"} Oct 14 09:20:21 crc kubenswrapper[5002]: I1014 09:20:21.454285 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g86rw" podStartSLOduration=2.855309739 podStartE2EDuration="6.454264895s" podCreationTimestamp="2025-10-14 09:20:15 +0000 UTC" firstStartedPulling="2025-10-14 09:20:17.390613241 +0000 UTC m=+5350.371852703" lastFinishedPulling="2025-10-14 09:20:20.989568407 +0000 UTC m=+5353.970807859" observedRunningTime="2025-10-14 09:20:21.453983947 +0000 UTC m=+5354.435223399" watchObservedRunningTime="2025-10-14 09:20:21.454264895 +0000 UTC m=+5354.435504347" Oct 14 09:20:26 crc kubenswrapper[5002]: I1014 09:20:26.073755 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:26 crc kubenswrapper[5002]: I1014 09:20:26.074552 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:27 crc kubenswrapper[5002]: I1014 09:20:27.124122 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-g86rw" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="registry-server" probeResult="failure" output=< Oct 14 09:20:27 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:20:27 crc kubenswrapper[5002]: > Oct 14 09:20:36 crc kubenswrapper[5002]: I1014 09:20:36.133468 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:36 crc kubenswrapper[5002]: I1014 09:20:36.202076 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:36 crc kubenswrapper[5002]: I1014 09:20:36.378732 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:37 crc kubenswrapper[5002]: I1014 09:20:37.585589 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g86rw" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="registry-server" containerID="cri-o://07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf" gracePeriod=2 Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.229709 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.404599 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content\") pod \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.404679 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities\") pod \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.404724 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf9hn\" (UniqueName: \"kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn\") pod \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\" (UID: \"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2\") " Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.405870 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities" (OuterVolumeSpecName: "utilities") pod "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" (UID: "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.412025 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn" (OuterVolumeSpecName: "kube-api-access-kf9hn") pod "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" (UID: "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2"). InnerVolumeSpecName "kube-api-access-kf9hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.455218 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" (UID: "7a12ceb9-2e5a-47f4-80d7-245c9b4464a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.507180 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.507211 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.507222 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf9hn\" (UniqueName: \"kubernetes.io/projected/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2-kube-api-access-kf9hn\") on node \"crc\" DevicePath \"\"" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.597689 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerID="07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf" exitCode=0 Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.597723 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerDied","Data":"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf"} Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.597798 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g86rw" event={"ID":"7a12ceb9-2e5a-47f4-80d7-245c9b4464a2","Type":"ContainerDied","Data":"b05da18433599035bf6cec4c6c2958664315baf9fb765b3423f0c348b2d446a8"} Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.597800 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g86rw" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.597826 5002 scope.go:117] "RemoveContainer" containerID="07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.636823 5002 scope.go:117] "RemoveContainer" containerID="17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.648551 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.659195 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g86rw"] Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.666887 5002 scope.go:117] "RemoveContainer" containerID="938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.718228 5002 scope.go:117] "RemoveContainer" containerID="07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf" Oct 14 09:20:38 crc kubenswrapper[5002]: E1014 09:20:38.718709 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf\": container with ID starting with 07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf not found: ID does not exist" containerID="07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.718751 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf"} err="failed to get container status \"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf\": rpc error: code = NotFound desc = could not find container \"07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf\": container with ID starting with 07e54d109bece42d7fc2424e905fd132d1dc4d541a8adf1231037f32d39cf2bf not found: ID does not exist" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.718777 5002 scope.go:117] "RemoveContainer" containerID="17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541" Oct 14 09:20:38 crc kubenswrapper[5002]: E1014 09:20:38.719200 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541\": container with ID starting with 17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541 not found: ID does not exist" containerID="17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.719227 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541"} err="failed to get container status \"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541\": rpc error: code = NotFound desc = could not find container \"17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541\": container with ID starting with 17ceb0dc5aea6a313bdd48aec9a2594610e30545f91e3cdd1488fa149cd4b541 not found: ID does not exist" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.719244 5002 scope.go:117] "RemoveContainer" containerID="938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034" Oct 14 09:20:38 crc kubenswrapper[5002]: E1014 09:20:38.719518 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034\": container with ID starting with 938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034 not found: ID does not exist" containerID="938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034" Oct 14 09:20:38 crc kubenswrapper[5002]: I1014 09:20:38.719545 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034"} err="failed to get container status \"938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034\": rpc error: code = NotFound desc = could not find container \"938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034\": container with ID starting with 938086048fb622c401fb8997de2bb614b80761bd592507332da0156ebcf66034 not found: ID does not exist" Oct 14 09:20:39 crc kubenswrapper[5002]: I1014 09:20:39.218958 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:20:39 crc kubenswrapper[5002]: I1014 09:20:39.219285 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:20:39 crc kubenswrapper[5002]: I1014 09:20:39.742329 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" path="/var/lib/kubelet/pods/7a12ceb9-2e5a-47f4-80d7-245c9b4464a2/volumes" Oct 14 09:21:09 crc kubenswrapper[5002]: I1014 09:21:09.218370 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:21:09 crc kubenswrapper[5002]: I1014 09:21:09.220323 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:21:39 crc kubenswrapper[5002]: I1014 09:21:39.218520 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:21:39 crc kubenswrapper[5002]: I1014 09:21:39.219089 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:21:39 crc kubenswrapper[5002]: I1014 09:21:39.219135 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:21:39 crc kubenswrapper[5002]: I1014 09:21:39.219877 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:21:39 crc kubenswrapper[5002]: I1014 09:21:39.219932 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185" gracePeriod=600 Oct 14 09:21:40 crc kubenswrapper[5002]: I1014 09:21:40.185441 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185" exitCode=0 Oct 14 09:21:40 crc kubenswrapper[5002]: I1014 09:21:40.185675 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185"} Oct 14 09:21:40 crc kubenswrapper[5002]: I1014 09:21:40.186019 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2"} Oct 14 09:21:40 crc kubenswrapper[5002]: I1014 09:21:40.186041 5002 scope.go:117] "RemoveContainer" containerID="c01419b0dd00ca92dc32b204f42612866552ccfed5fc6df1191a1216648551e5" Oct 14 09:23:39 crc kubenswrapper[5002]: I1014 09:23:39.218411 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:23:39 crc kubenswrapper[5002]: I1014 09:23:39.218969 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:24:09 crc kubenswrapper[5002]: I1014 09:24:09.221070 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:24:09 crc kubenswrapper[5002]: I1014 09:24:09.221665 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:24:39 crc kubenswrapper[5002]: I1014 09:24:39.217960 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:24:39 crc kubenswrapper[5002]: I1014 09:24:39.218579 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:24:39 crc kubenswrapper[5002]: I1014 09:24:39.218639 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:24:39 crc kubenswrapper[5002]: I1014 09:24:39.219645 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:24:39 crc kubenswrapper[5002]: I1014 09:24:39.219722 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" gracePeriod=600 Oct 14 09:24:39 crc kubenswrapper[5002]: E1014 09:24:39.351338 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:24:40 crc kubenswrapper[5002]: I1014 09:24:40.006652 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" exitCode=0 Oct 14 09:24:40 crc kubenswrapper[5002]: I1014 09:24:40.006734 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2"} Oct 14 09:24:40 crc kubenswrapper[5002]: I1014 09:24:40.006809 5002 scope.go:117] "RemoveContainer" containerID="5290df7d547b236a4869792b42b173ceaefc33aa81f926fb7c0342e21f34e185" Oct 14 09:24:40 crc kubenswrapper[5002]: I1014 09:24:40.008122 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:24:40 crc kubenswrapper[5002]: E1014 09:24:40.008727 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:24:54 crc kubenswrapper[5002]: I1014 09:24:54.721710 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:24:54 crc kubenswrapper[5002]: E1014 09:24:54.723901 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:25:06 crc kubenswrapper[5002]: I1014 09:25:06.720899 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:25:06 crc kubenswrapper[5002]: E1014 09:25:06.721608 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:25:21 crc kubenswrapper[5002]: I1014 09:25:21.727943 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:25:21 crc kubenswrapper[5002]: E1014 09:25:21.728782 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:25:34 crc kubenswrapper[5002]: I1014 09:25:34.722076 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:25:34 crc kubenswrapper[5002]: E1014 09:25:34.723433 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:25:49 crc kubenswrapper[5002]: I1014 09:25:49.720609 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:25:49 crc kubenswrapper[5002]: E1014 09:25:49.723149 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:26:03 crc kubenswrapper[5002]: I1014 09:26:03.720900 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:26:03 crc kubenswrapper[5002]: E1014 09:26:03.721723 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:26:14 crc kubenswrapper[5002]: I1014 09:26:14.721461 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:26:14 crc kubenswrapper[5002]: E1014 09:26:14.722266 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:26:25 crc kubenswrapper[5002]: I1014 09:26:25.721334 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:26:25 crc kubenswrapper[5002]: E1014 09:26:25.722650 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:26:36 crc kubenswrapper[5002]: I1014 09:26:36.721014 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:26:36 crc kubenswrapper[5002]: E1014 09:26:36.721897 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:26:48 crc kubenswrapper[5002]: I1014 09:26:48.720042 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:26:48 crc kubenswrapper[5002]: E1014 09:26:48.720761 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:27:02 crc kubenswrapper[5002]: I1014 09:27:02.721469 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:27:02 crc kubenswrapper[5002]: E1014 09:27:02.722507 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:27:15 crc kubenswrapper[5002]: I1014 09:27:15.720428 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:27:15 crc kubenswrapper[5002]: E1014 09:27:15.722423 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:27:28 crc kubenswrapper[5002]: I1014 09:27:28.720738 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:27:28 crc kubenswrapper[5002]: E1014 09:27:28.721651 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:27:39 crc kubenswrapper[5002]: I1014 09:27:39.721244 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:27:39 crc kubenswrapper[5002]: E1014 09:27:39.722073 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:27:52 crc kubenswrapper[5002]: I1014 09:27:52.720491 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:27:52 crc kubenswrapper[5002]: E1014 09:27:52.721270 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:28:05 crc kubenswrapper[5002]: I1014 09:28:05.721491 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:28:05 crc kubenswrapper[5002]: E1014 09:28:05.722579 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:28:16 crc kubenswrapper[5002]: I1014 09:28:16.720862 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:28:16 crc kubenswrapper[5002]: E1014 09:28:16.721751 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:28:28 crc kubenswrapper[5002]: I1014 09:28:28.721203 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:28:28 crc kubenswrapper[5002]: E1014 09:28:28.722112 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:28:40 crc kubenswrapper[5002]: I1014 09:28:40.720384 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:28:40 crc kubenswrapper[5002]: E1014 09:28:40.721672 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:28:54 crc kubenswrapper[5002]: I1014 09:28:54.721061 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:28:54 crc kubenswrapper[5002]: E1014 09:28:54.721761 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:29:05 crc kubenswrapper[5002]: I1014 09:29:05.721647 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:29:05 crc kubenswrapper[5002]: E1014 09:29:05.722690 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:29:16 crc kubenswrapper[5002]: I1014 09:29:16.720536 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:29:16 crc kubenswrapper[5002]: E1014 09:29:16.721593 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.312779 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:26 crc kubenswrapper[5002]: E1014 09:29:26.314498 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="extract-utilities" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.314541 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="extract-utilities" Oct 14 09:29:26 crc kubenswrapper[5002]: E1014 09:29:26.314581 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="registry-server" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.314601 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="registry-server" Oct 14 09:29:26 crc kubenswrapper[5002]: E1014 09:29:26.314658 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="extract-content" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.314681 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="extract-content" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.315505 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a12ceb9-2e5a-47f4-80d7-245c9b4464a2" containerName="registry-server" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.319012 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.353232 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.382343 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5szw\" (UniqueName: \"kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.382461 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.382552 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.484647 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5szw\" (UniqueName: \"kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.484744 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.484813 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.485336 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.485414 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.507432 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5szw\" (UniqueName: \"kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw\") pod \"redhat-operators-kndht\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:26 crc kubenswrapper[5002]: I1014 09:29:26.663489 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:27 crc kubenswrapper[5002]: I1014 09:29:27.119965 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:27 crc kubenswrapper[5002]: I1014 09:29:27.941111 5002 generic.go:334] "Generic (PLEG): container finished" podID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerID="ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e" exitCode=0 Oct 14 09:29:27 crc kubenswrapper[5002]: I1014 09:29:27.941191 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerDied","Data":"ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e"} Oct 14 09:29:27 crc kubenswrapper[5002]: I1014 09:29:27.941373 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerStarted","Data":"b206ef1a7d39dfdb352efe436e82f8ba7c37afcf2bac23558cfb95ef194f0a9c"} Oct 14 09:29:27 crc kubenswrapper[5002]: I1014 09:29:27.944668 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:29:29 crc kubenswrapper[5002]: I1014 09:29:29.975087 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerStarted","Data":"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96"} Oct 14 09:29:30 crc kubenswrapper[5002]: I1014 09:29:30.720776 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:29:30 crc kubenswrapper[5002]: E1014 09:29:30.721309 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:29:33 crc kubenswrapper[5002]: I1014 09:29:33.012011 5002 generic.go:334] "Generic (PLEG): container finished" podID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerID="ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96" exitCode=0 Oct 14 09:29:33 crc kubenswrapper[5002]: I1014 09:29:33.012937 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerDied","Data":"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96"} Oct 14 09:29:34 crc kubenswrapper[5002]: I1014 09:29:34.025478 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerStarted","Data":"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d"} Oct 14 09:29:34 crc kubenswrapper[5002]: I1014 09:29:34.053395 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kndht" podStartSLOduration=2.490099172 podStartE2EDuration="8.05337373s" podCreationTimestamp="2025-10-14 09:29:26 +0000 UTC" firstStartedPulling="2025-10-14 09:29:27.944134992 +0000 UTC m=+5900.925374484" lastFinishedPulling="2025-10-14 09:29:33.50740958 +0000 UTC m=+5906.488649042" observedRunningTime="2025-10-14 09:29:34.04438834 +0000 UTC m=+5907.025627842" watchObservedRunningTime="2025-10-14 09:29:34.05337373 +0000 UTC m=+5907.034613192" Oct 14 09:29:36 crc kubenswrapper[5002]: I1014 09:29:36.664364 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:36 crc kubenswrapper[5002]: I1014 09:29:36.664664 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:37 crc kubenswrapper[5002]: I1014 09:29:37.730759 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kndht" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" probeResult="failure" output=< Oct 14 09:29:37 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:29:37 crc kubenswrapper[5002]: > Oct 14 09:29:42 crc kubenswrapper[5002]: I1014 09:29:42.720994 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:29:43 crc kubenswrapper[5002]: I1014 09:29:43.113273 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75"} Oct 14 09:29:47 crc kubenswrapper[5002]: I1014 09:29:47.733314 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kndht" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" probeResult="failure" output=< Oct 14 09:29:47 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:29:47 crc kubenswrapper[5002]: > Oct 14 09:29:56 crc kubenswrapper[5002]: I1014 09:29:56.710634 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:56 crc kubenswrapper[5002]: I1014 09:29:56.786288 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:57 crc kubenswrapper[5002]: I1014 09:29:57.505170 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.294222 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kndht" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" containerID="cri-o://2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d" gracePeriod=2 Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.860103 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.976006 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5szw\" (UniqueName: \"kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw\") pod \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.976351 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities\") pod \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.976517 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content\") pod \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\" (UID: \"1b4c9d0a-30b1-4db9-86c5-1966300f55d5\") " Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.976950 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities" (OuterVolumeSpecName: "utilities") pod "1b4c9d0a-30b1-4db9-86c5-1966300f55d5" (UID: "1b4c9d0a-30b1-4db9-86c5-1966300f55d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.977435 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:58 crc kubenswrapper[5002]: I1014 09:29:58.986564 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw" (OuterVolumeSpecName: "kube-api-access-v5szw") pod "1b4c9d0a-30b1-4db9-86c5-1966300f55d5" (UID: "1b4c9d0a-30b1-4db9-86c5-1966300f55d5"). InnerVolumeSpecName "kube-api-access-v5szw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.060599 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b4c9d0a-30b1-4db9-86c5-1966300f55d5" (UID: "1b4c9d0a-30b1-4db9-86c5-1966300f55d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.078990 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.079032 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5szw\" (UniqueName: \"kubernetes.io/projected/1b4c9d0a-30b1-4db9-86c5-1966300f55d5-kube-api-access-v5szw\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.311362 5002 generic.go:334] "Generic (PLEG): container finished" podID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerID="2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d" exitCode=0 Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.311427 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerDied","Data":"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d"} Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.311470 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kndht" event={"ID":"1b4c9d0a-30b1-4db9-86c5-1966300f55d5","Type":"ContainerDied","Data":"b206ef1a7d39dfdb352efe436e82f8ba7c37afcf2bac23558cfb95ef194f0a9c"} Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.311470 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kndht" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.311499 5002 scope.go:117] "RemoveContainer" containerID="2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.348690 5002 scope.go:117] "RemoveContainer" containerID="ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.385334 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.396348 5002 scope.go:117] "RemoveContainer" containerID="ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.400221 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kndht"] Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.458935 5002 scope.go:117] "RemoveContainer" containerID="2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d" Oct 14 09:29:59 crc kubenswrapper[5002]: E1014 09:29:59.459609 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d\": container with ID starting with 2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d not found: ID does not exist" containerID="2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.459653 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d"} err="failed to get container status \"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d\": rpc error: code = NotFound desc = could not find container \"2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d\": container with ID starting with 2333bb6dcf7c608fe7511c27129c31d1ebfdf40fccb31d9b5071184b8d4e595d not found: ID does not exist" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.459681 5002 scope.go:117] "RemoveContainer" containerID="ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96" Oct 14 09:29:59 crc kubenswrapper[5002]: E1014 09:29:59.460362 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96\": container with ID starting with ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96 not found: ID does not exist" containerID="ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.460411 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96"} err="failed to get container status \"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96\": rpc error: code = NotFound desc = could not find container \"ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96\": container with ID starting with ebf711dfb5c0010c003a0c720f5b73bfcba2061e453783ff2387b5448d83fe96 not found: ID does not exist" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.460445 5002 scope.go:117] "RemoveContainer" containerID="ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e" Oct 14 09:29:59 crc kubenswrapper[5002]: E1014 09:29:59.460770 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e\": container with ID starting with ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e not found: ID does not exist" containerID="ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.460804 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e"} err="failed to get container status \"ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e\": rpc error: code = NotFound desc = could not find container \"ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e\": container with ID starting with ebdccacb8deae5620956873ea5376dd0dea57f3f12f31320a7d28e6a28c5b89e not found: ID does not exist" Oct 14 09:29:59 crc kubenswrapper[5002]: I1014 09:29:59.743160 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" path="/var/lib/kubelet/pods/1b4c9d0a-30b1-4db9-86c5-1966300f55d5/volumes" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.165134 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7"] Oct 14 09:30:00 crc kubenswrapper[5002]: E1014 09:30:00.168140 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.168170 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[5002]: E1014 09:30:00.168194 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="extract-utilities" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.168203 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="extract-utilities" Oct 14 09:30:00 crc kubenswrapper[5002]: E1014 09:30:00.168212 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="extract-content" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.168220 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="extract-content" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.168491 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b4c9d0a-30b1-4db9-86c5-1966300f55d5" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.169425 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.174267 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.174977 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.187564 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7"] Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.308791 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.308870 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.309006 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lf6\" (UniqueName: \"kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.410278 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lf6\" (UniqueName: \"kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.410387 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.410410 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.413498 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.422278 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.426179 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lf6\" (UniqueName: \"kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6\") pod \"collect-profiles-29340570-27lr7\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.503280 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:00 crc kubenswrapper[5002]: I1014 09:30:00.990074 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7"] Oct 14 09:30:01 crc kubenswrapper[5002]: I1014 09:30:01.332660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" event={"ID":"344fc1f0-aa77-4cb2-bd81-9390399e0b23","Type":"ContainerStarted","Data":"99577c3afad10959cde12113b5ebcd84e03ff4cffc57a54be501fcaf4afd9482"} Oct 14 09:30:01 crc kubenswrapper[5002]: I1014 09:30:01.333019 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" event={"ID":"344fc1f0-aa77-4cb2-bd81-9390399e0b23","Type":"ContainerStarted","Data":"a9eec3e1d68c7e228950606352b7979c17d711acfd1331a1c9ba63496c817065"} Oct 14 09:30:01 crc kubenswrapper[5002]: I1014 09:30:01.356927 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" podStartSLOduration=1.356907064 podStartE2EDuration="1.356907064s" podCreationTimestamp="2025-10-14 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:30:01.346399743 +0000 UTC m=+5934.327639225" watchObservedRunningTime="2025-10-14 09:30:01.356907064 +0000 UTC m=+5934.338146516" Oct 14 09:30:02 crc kubenswrapper[5002]: I1014 09:30:02.357175 5002 generic.go:334] "Generic (PLEG): container finished" podID="344fc1f0-aa77-4cb2-bd81-9390399e0b23" containerID="99577c3afad10959cde12113b5ebcd84e03ff4cffc57a54be501fcaf4afd9482" exitCode=0 Oct 14 09:30:02 crc kubenswrapper[5002]: I1014 09:30:02.357788 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" event={"ID":"344fc1f0-aa77-4cb2-bd81-9390399e0b23","Type":"ContainerDied","Data":"99577c3afad10959cde12113b5ebcd84e03ff4cffc57a54be501fcaf4afd9482"} Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.809770 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.885781 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume\") pod \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.885903 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9lf6\" (UniqueName: \"kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6\") pod \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.886125 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume\") pod \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\" (UID: \"344fc1f0-aa77-4cb2-bd81-9390399e0b23\") " Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.888552 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume" (OuterVolumeSpecName: "config-volume") pod "344fc1f0-aa77-4cb2-bd81-9390399e0b23" (UID: "344fc1f0-aa77-4cb2-bd81-9390399e0b23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.893819 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "344fc1f0-aa77-4cb2-bd81-9390399e0b23" (UID: "344fc1f0-aa77-4cb2-bd81-9390399e0b23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.894389 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6" (OuterVolumeSpecName: "kube-api-access-t9lf6") pod "344fc1f0-aa77-4cb2-bd81-9390399e0b23" (UID: "344fc1f0-aa77-4cb2-bd81-9390399e0b23"). InnerVolumeSpecName "kube-api-access-t9lf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.988315 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9lf6\" (UniqueName: \"kubernetes.io/projected/344fc1f0-aa77-4cb2-bd81-9390399e0b23-kube-api-access-t9lf6\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.988853 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/344fc1f0-aa77-4cb2-bd81-9390399e0b23-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:03 crc kubenswrapper[5002]: I1014 09:30:03.988931 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/344fc1f0-aa77-4cb2-bd81-9390399e0b23-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:04 crc kubenswrapper[5002]: I1014 09:30:04.375507 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" event={"ID":"344fc1f0-aa77-4cb2-bd81-9390399e0b23","Type":"ContainerDied","Data":"a9eec3e1d68c7e228950606352b7979c17d711acfd1331a1c9ba63496c817065"} Oct 14 09:30:04 crc kubenswrapper[5002]: I1014 09:30:04.375550 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9eec3e1d68c7e228950606352b7979c17d711acfd1331a1c9ba63496c817065" Oct 14 09:30:04 crc kubenswrapper[5002]: I1014 09:30:04.375589 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-27lr7" Oct 14 09:30:04 crc kubenswrapper[5002]: I1014 09:30:04.423143 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm"] Oct 14 09:30:04 crc kubenswrapper[5002]: I1014 09:30:04.430742 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-tr2tm"] Oct 14 09:30:05 crc kubenswrapper[5002]: I1014 09:30:05.738640 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="853cca8b-4f20-40a5-91ae-50e50417395f" path="/var/lib/kubelet/pods/853cca8b-4f20-40a5-91ae-50e50417395f/volumes" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.071323 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:06 crc kubenswrapper[5002]: E1014 09:30:06.072756 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="344fc1f0-aa77-4cb2-bd81-9390399e0b23" containerName="collect-profiles" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.072788 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="344fc1f0-aa77-4cb2-bd81-9390399e0b23" containerName="collect-profiles" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.073733 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="344fc1f0-aa77-4cb2-bd81-9390399e0b23" containerName="collect-profiles" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.104781 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.104946 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.234784 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.234868 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.234911 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txrxm\" (UniqueName: \"kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.336860 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.336954 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txrxm\" (UniqueName: \"kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.337110 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.337282 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.337535 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.357566 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txrxm\" (UniqueName: \"kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm\") pod \"community-operators-n7zpd\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.438617 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:06 crc kubenswrapper[5002]: I1014 09:30:06.939183 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:07 crc kubenswrapper[5002]: I1014 09:30:07.407548 5002 generic.go:334] "Generic (PLEG): container finished" podID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerID="f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de" exitCode=0 Oct 14 09:30:07 crc kubenswrapper[5002]: I1014 09:30:07.407660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerDied","Data":"f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de"} Oct 14 09:30:07 crc kubenswrapper[5002]: I1014 09:30:07.408021 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerStarted","Data":"0ae651dd7d32b909cab9a210c0fa8178028ebec7d3cb85b7b97a19e202b8a64c"} Oct 14 09:30:08 crc kubenswrapper[5002]: I1014 09:30:08.422601 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerStarted","Data":"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5"} Oct 14 09:30:10 crc kubenswrapper[5002]: I1014 09:30:10.461347 5002 generic.go:334] "Generic (PLEG): container finished" podID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerID="a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5" exitCode=0 Oct 14 09:30:10 crc kubenswrapper[5002]: I1014 09:30:10.461432 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerDied","Data":"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5"} Oct 14 09:30:11 crc kubenswrapper[5002]: I1014 09:30:11.476009 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerStarted","Data":"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b"} Oct 14 09:30:11 crc kubenswrapper[5002]: I1014 09:30:11.500951 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n7zpd" podStartSLOduration=1.863215012 podStartE2EDuration="5.500925325s" podCreationTimestamp="2025-10-14 09:30:06 +0000 UTC" firstStartedPulling="2025-10-14 09:30:07.409878368 +0000 UTC m=+5940.391117900" lastFinishedPulling="2025-10-14 09:30:11.047588721 +0000 UTC m=+5944.028828213" observedRunningTime="2025-10-14 09:30:11.497596045 +0000 UTC m=+5944.478835527" watchObservedRunningTime="2025-10-14 09:30:11.500925325 +0000 UTC m=+5944.482164817" Oct 14 09:30:16 crc kubenswrapper[5002]: I1014 09:30:16.439905 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:16 crc kubenswrapper[5002]: I1014 09:30:16.440781 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:16 crc kubenswrapper[5002]: I1014 09:30:16.515046 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:16 crc kubenswrapper[5002]: I1014 09:30:16.595678 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:16 crc kubenswrapper[5002]: I1014 09:30:16.771111 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:18 crc kubenswrapper[5002]: I1014 09:30:18.544026 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n7zpd" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="registry-server" containerID="cri-o://b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b" gracePeriod=2 Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.157086 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.225682 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txrxm\" (UniqueName: \"kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm\") pod \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.225746 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities\") pod \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.225984 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content\") pod \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\" (UID: \"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f\") " Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.226931 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities" (OuterVolumeSpecName: "utilities") pod "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" (UID: "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.235076 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm" (OuterVolumeSpecName: "kube-api-access-txrxm") pod "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" (UID: "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f"). InnerVolumeSpecName "kube-api-access-txrxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.273881 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" (UID: "f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.328911 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.329248 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txrxm\" (UniqueName: \"kubernetes.io/projected/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-kube-api-access-txrxm\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.329451 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.558978 5002 generic.go:334] "Generic (PLEG): container finished" podID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerID="b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b" exitCode=0 Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.559065 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerDied","Data":"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b"} Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.559097 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n7zpd" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.559546 5002 scope.go:117] "RemoveContainer" containerID="b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.560136 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n7zpd" event={"ID":"f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f","Type":"ContainerDied","Data":"0ae651dd7d32b909cab9a210c0fa8178028ebec7d3cb85b7b97a19e202b8a64c"} Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.593127 5002 scope.go:117] "RemoveContainer" containerID="a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.625391 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.637610 5002 scope.go:117] "RemoveContainer" containerID="f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.643052 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n7zpd"] Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.699203 5002 scope.go:117] "RemoveContainer" containerID="b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b" Oct 14 09:30:19 crc kubenswrapper[5002]: E1014 09:30:19.699886 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b\": container with ID starting with b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b not found: ID does not exist" containerID="b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.700103 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b"} err="failed to get container status \"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b\": rpc error: code = NotFound desc = could not find container \"b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b\": container with ID starting with b0969021839e730ae451af50e6273a52e7586f4eb4d95467ab7facd0ef5e943b not found: ID does not exist" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.700286 5002 scope.go:117] "RemoveContainer" containerID="a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5" Oct 14 09:30:19 crc kubenswrapper[5002]: E1014 09:30:19.700971 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5\": container with ID starting with a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5 not found: ID does not exist" containerID="a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.701022 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5"} err="failed to get container status \"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5\": rpc error: code = NotFound desc = could not find container \"a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5\": container with ID starting with a5da76224ae58f723290a156edeb9ac720c91b470b6d0fcf45330efbc9e17da5 not found: ID does not exist" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.701057 5002 scope.go:117] "RemoveContainer" containerID="f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de" Oct 14 09:30:19 crc kubenswrapper[5002]: E1014 09:30:19.701440 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de\": container with ID starting with f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de not found: ID does not exist" containerID="f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.701641 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de"} err="failed to get container status \"f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de\": rpc error: code = NotFound desc = could not find container \"f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de\": container with ID starting with f23b5c27c745b3a8f97ccc17f1fb6c0dd9b5939ecc9a6ccd6354693226df86de not found: ID does not exist" Oct 14 09:30:19 crc kubenswrapper[5002]: I1014 09:30:19.745727 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" path="/var/lib/kubelet/pods/f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f/volumes" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.110924 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:25 crc kubenswrapper[5002]: E1014 09:30:25.112923 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="registry-server" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.112944 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="registry-server" Oct 14 09:30:25 crc kubenswrapper[5002]: E1014 09:30:25.112970 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="extract-content" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.112978 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="extract-content" Oct 14 09:30:25 crc kubenswrapper[5002]: E1014 09:30:25.113008 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="extract-utilities" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.113017 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="extract-utilities" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.113255 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f436cb28-0ca2-42a3-bf0a-7c4b1d046b2f" containerName="registry-server" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.114987 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.131200 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.263184 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.263532 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fks2f\" (UniqueName: \"kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.263726 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.365938 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.366046 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fks2f\" (UniqueName: \"kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.366117 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.366771 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.367016 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.447787 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fks2f\" (UniqueName: \"kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f\") pod \"redhat-marketplace-h9chd\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.462925 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:25 crc kubenswrapper[5002]: I1014 09:30:25.951782 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:25 crc kubenswrapper[5002]: W1014 09:30:25.959968 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c7d3c63_3c1c_4c17_9a48_91753cc29e4d.slice/crio-c1ca992a6223ab3f75d881425cc6f2c813cc85c7e6a0cdc7e18265a0d3703a1e WatchSource:0}: Error finding container c1ca992a6223ab3f75d881425cc6f2c813cc85c7e6a0cdc7e18265a0d3703a1e: Status 404 returned error can't find the container with id c1ca992a6223ab3f75d881425cc6f2c813cc85c7e6a0cdc7e18265a0d3703a1e Oct 14 09:30:26 crc kubenswrapper[5002]: I1014 09:30:26.631459 5002 generic.go:334] "Generic (PLEG): container finished" podID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerID="94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472" exitCode=0 Oct 14 09:30:26 crc kubenswrapper[5002]: I1014 09:30:26.631529 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerDied","Data":"94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472"} Oct 14 09:30:26 crc kubenswrapper[5002]: I1014 09:30:26.631736 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerStarted","Data":"c1ca992a6223ab3f75d881425cc6f2c813cc85c7e6a0cdc7e18265a0d3703a1e"} Oct 14 09:30:28 crc kubenswrapper[5002]: I1014 09:30:28.649394 5002 generic.go:334] "Generic (PLEG): container finished" podID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerID="ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be" exitCode=0 Oct 14 09:30:28 crc kubenswrapper[5002]: I1014 09:30:28.649465 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerDied","Data":"ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be"} Oct 14 09:30:29 crc kubenswrapper[5002]: I1014 09:30:29.661455 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerStarted","Data":"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae"} Oct 14 09:30:29 crc kubenswrapper[5002]: I1014 09:30:29.695052 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h9chd" podStartSLOduration=2.31319077 podStartE2EDuration="4.695029657s" podCreationTimestamp="2025-10-14 09:30:25 +0000 UTC" firstStartedPulling="2025-10-14 09:30:26.634772578 +0000 UTC m=+5959.616012030" lastFinishedPulling="2025-10-14 09:30:29.016611445 +0000 UTC m=+5961.997850917" observedRunningTime="2025-10-14 09:30:29.684343212 +0000 UTC m=+5962.665582674" watchObservedRunningTime="2025-10-14 09:30:29.695029657 +0000 UTC m=+5962.676269129" Oct 14 09:30:30 crc kubenswrapper[5002]: I1014 09:30:30.409092 5002 scope.go:117] "RemoveContainer" containerID="6391c8c9491417a713d8da78dbe3e379d8d87aba5a3f1f23b631364ee5953b1b" Oct 14 09:30:31 crc kubenswrapper[5002]: I1014 09:30:31.915263 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:31 crc kubenswrapper[5002]: I1014 09:30:31.919398 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:31 crc kubenswrapper[5002]: I1014 09:30:31.932260 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.025913 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.026224 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.026312 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95z5l\" (UniqueName: \"kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.128355 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95z5l\" (UniqueName: \"kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.128499 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.128593 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.129158 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.129437 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.152350 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95z5l\" (UniqueName: \"kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l\") pod \"certified-operators-gm8vx\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.258284 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:32 crc kubenswrapper[5002]: I1014 09:30:32.827979 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:33 crc kubenswrapper[5002]: I1014 09:30:33.708581 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerID="3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45" exitCode=0 Oct 14 09:30:33 crc kubenswrapper[5002]: I1014 09:30:33.708897 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerDied","Data":"3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45"} Oct 14 09:30:33 crc kubenswrapper[5002]: I1014 09:30:33.708955 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerStarted","Data":"e5522ac7a8d54ef20ca2616e84393ee5f8e65a8177f9aa6d20f8c35481f4fbd5"} Oct 14 09:30:35 crc kubenswrapper[5002]: I1014 09:30:35.463308 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:35 crc kubenswrapper[5002]: I1014 09:30:35.464020 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:35 crc kubenswrapper[5002]: I1014 09:30:35.523303 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:35 crc kubenswrapper[5002]: I1014 09:30:35.732576 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerStarted","Data":"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693"} Oct 14 09:30:35 crc kubenswrapper[5002]: I1014 09:30:35.793928 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:36 crc kubenswrapper[5002]: I1014 09:30:36.741280 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerID="9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693" exitCode=0 Oct 14 09:30:36 crc kubenswrapper[5002]: I1014 09:30:36.741420 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerDied","Data":"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693"} Oct 14 09:30:37 crc kubenswrapper[5002]: I1014 09:30:37.483492 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:37 crc kubenswrapper[5002]: I1014 09:30:37.750603 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerStarted","Data":"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068"} Oct 14 09:30:37 crc kubenswrapper[5002]: I1014 09:30:37.750748 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h9chd" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="registry-server" containerID="cri-o://cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae" gracePeriod=2 Oct 14 09:30:37 crc kubenswrapper[5002]: I1014 09:30:37.778343 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gm8vx" podStartSLOduration=3.203096006 podStartE2EDuration="6.778323788s" podCreationTimestamp="2025-10-14 09:30:31 +0000 UTC" firstStartedPulling="2025-10-14 09:30:33.711801718 +0000 UTC m=+5966.693041170" lastFinishedPulling="2025-10-14 09:30:37.2870295 +0000 UTC m=+5970.268268952" observedRunningTime="2025-10-14 09:30:37.775473182 +0000 UTC m=+5970.756712634" watchObservedRunningTime="2025-10-14 09:30:37.778323788 +0000 UTC m=+5970.759563250" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.369092 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.499724 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities\") pod \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.499879 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fks2f\" (UniqueName: \"kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f\") pod \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.499972 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content\") pod \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\" (UID: \"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d\") " Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.500724 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities" (OuterVolumeSpecName: "utilities") pod "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" (UID: "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.507105 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f" (OuterVolumeSpecName: "kube-api-access-fks2f") pod "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" (UID: "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d"). InnerVolumeSpecName "kube-api-access-fks2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.511978 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" (UID: "3c7d3c63-3c1c-4c17-9a48-91753cc29e4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.602901 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.603186 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fks2f\" (UniqueName: \"kubernetes.io/projected/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-kube-api-access-fks2f\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.603198 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.763720 5002 generic.go:334] "Generic (PLEG): container finished" podID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerID="cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae" exitCode=0 Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.763809 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h9chd" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.763813 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerDied","Data":"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae"} Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.764075 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h9chd" event={"ID":"3c7d3c63-3c1c-4c17-9a48-91753cc29e4d","Type":"ContainerDied","Data":"c1ca992a6223ab3f75d881425cc6f2c813cc85c7e6a0cdc7e18265a0d3703a1e"} Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.764117 5002 scope.go:117] "RemoveContainer" containerID="cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.795703 5002 scope.go:117] "RemoveContainer" containerID="ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.814825 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.825985 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h9chd"] Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.829913 5002 scope.go:117] "RemoveContainer" containerID="94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.887378 5002 scope.go:117] "RemoveContainer" containerID="cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae" Oct 14 09:30:38 crc kubenswrapper[5002]: E1014 09:30:38.887979 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae\": container with ID starting with cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae not found: ID does not exist" containerID="cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.888032 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae"} err="failed to get container status \"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae\": rpc error: code = NotFound desc = could not find container \"cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae\": container with ID starting with cd21d7b4698eb0f92154936c1c33a427bb08fdf01204f868c902212809549dae not found: ID does not exist" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.888057 5002 scope.go:117] "RemoveContainer" containerID="ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be" Oct 14 09:30:38 crc kubenswrapper[5002]: E1014 09:30:38.888653 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be\": container with ID starting with ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be not found: ID does not exist" containerID="ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.888704 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be"} err="failed to get container status \"ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be\": rpc error: code = NotFound desc = could not find container \"ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be\": container with ID starting with ddd531babcdb4828cfe9471e33060d223500ee1f57a35eb314d5fbcdf0f6f7be not found: ID does not exist" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.888742 5002 scope.go:117] "RemoveContainer" containerID="94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472" Oct 14 09:30:38 crc kubenswrapper[5002]: E1014 09:30:38.889153 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472\": container with ID starting with 94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472 not found: ID does not exist" containerID="94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472" Oct 14 09:30:38 crc kubenswrapper[5002]: I1014 09:30:38.889200 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472"} err="failed to get container status \"94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472\": rpc error: code = NotFound desc = could not find container \"94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472\": container with ID starting with 94b4d715031ce37a5090a177384cd0ea28c108ff151ce3b5f67e4a1457b99472 not found: ID does not exist" Oct 14 09:30:39 crc kubenswrapper[5002]: I1014 09:30:39.730910 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" path="/var/lib/kubelet/pods/3c7d3c63-3c1c-4c17-9a48-91753cc29e4d/volumes" Oct 14 09:30:42 crc kubenswrapper[5002]: I1014 09:30:42.259870 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:42 crc kubenswrapper[5002]: I1014 09:30:42.260557 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:42 crc kubenswrapper[5002]: I1014 09:30:42.323795 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:42 crc kubenswrapper[5002]: I1014 09:30:42.869725 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:43 crc kubenswrapper[5002]: I1014 09:30:43.881412 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:44 crc kubenswrapper[5002]: I1014 09:30:44.835071 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gm8vx" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="registry-server" containerID="cri-o://45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068" gracePeriod=2 Oct 14 09:30:44 crc kubenswrapper[5002]: E1014 09:30:44.970398 5002 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce05640d_53b3_4703_9b19_ea9b67b76aa5.slice/crio-conmon-45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce05640d_53b3_4703_9b19_ea9b67b76aa5.slice/crio-45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.387205 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.563826 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95z5l\" (UniqueName: \"kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l\") pod \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.564320 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content\") pod \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.564381 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities\") pod \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\" (UID: \"ce05640d-53b3-4703-9b19-ea9b67b76aa5\") " Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.565345 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities" (OuterVolumeSpecName: "utilities") pod "ce05640d-53b3-4703-9b19-ea9b67b76aa5" (UID: "ce05640d-53b3-4703-9b19-ea9b67b76aa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.571139 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l" (OuterVolumeSpecName: "kube-api-access-95z5l") pod "ce05640d-53b3-4703-9b19-ea9b67b76aa5" (UID: "ce05640d-53b3-4703-9b19-ea9b67b76aa5"). InnerVolumeSpecName "kube-api-access-95z5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.619438 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce05640d-53b3-4703-9b19-ea9b67b76aa5" (UID: "ce05640d-53b3-4703-9b19-ea9b67b76aa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.666938 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.666987 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95z5l\" (UniqueName: \"kubernetes.io/projected/ce05640d-53b3-4703-9b19-ea9b67b76aa5-kube-api-access-95z5l\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.667009 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce05640d-53b3-4703-9b19-ea9b67b76aa5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.848085 5002 generic.go:334] "Generic (PLEG): container finished" podID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerID="45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068" exitCode=0 Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.848163 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerDied","Data":"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068"} Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.848222 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gm8vx" event={"ID":"ce05640d-53b3-4703-9b19-ea9b67b76aa5","Type":"ContainerDied","Data":"e5522ac7a8d54ef20ca2616e84393ee5f8e65a8177f9aa6d20f8c35481f4fbd5"} Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.848261 5002 scope.go:117] "RemoveContainer" containerID="45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.848475 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gm8vx" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.884848 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.890075 5002 scope.go:117] "RemoveContainer" containerID="9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.895616 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gm8vx"] Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.937728 5002 scope.go:117] "RemoveContainer" containerID="3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.991542 5002 scope.go:117] "RemoveContainer" containerID="45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068" Oct 14 09:30:45 crc kubenswrapper[5002]: E1014 09:30:45.992141 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068\": container with ID starting with 45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068 not found: ID does not exist" containerID="45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.992181 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068"} err="failed to get container status \"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068\": rpc error: code = NotFound desc = could not find container \"45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068\": container with ID starting with 45e2369c5891a610d1bc4041371539a23c05d66af3da88784595dcf780810068 not found: ID does not exist" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.992205 5002 scope.go:117] "RemoveContainer" containerID="9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693" Oct 14 09:30:45 crc kubenswrapper[5002]: E1014 09:30:45.992705 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693\": container with ID starting with 9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693 not found: ID does not exist" containerID="9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.992748 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693"} err="failed to get container status \"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693\": rpc error: code = NotFound desc = could not find container \"9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693\": container with ID starting with 9a1817e21e303332b53aaaad78fe7ddc57ea7f4011e43e3026bf4fd17a11b693 not found: ID does not exist" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.992777 5002 scope.go:117] "RemoveContainer" containerID="3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45" Oct 14 09:30:45 crc kubenswrapper[5002]: E1014 09:30:45.993264 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45\": container with ID starting with 3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45 not found: ID does not exist" containerID="3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45" Oct 14 09:30:45 crc kubenswrapper[5002]: I1014 09:30:45.993289 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45"} err="failed to get container status \"3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45\": rpc error: code = NotFound desc = could not find container \"3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45\": container with ID starting with 3bf385f90b7304d40429383fc434cac005a2d48675ae2a482d3e46bb54b89a45 not found: ID does not exist" Oct 14 09:30:47 crc kubenswrapper[5002]: I1014 09:30:47.740037 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" path="/var/lib/kubelet/pods/ce05640d-53b3-4703-9b19-ea9b67b76aa5/volumes" Oct 14 09:32:09 crc kubenswrapper[5002]: I1014 09:32:09.217966 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:32:09 crc kubenswrapper[5002]: I1014 09:32:09.218627 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:32:39 crc kubenswrapper[5002]: I1014 09:32:39.217931 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:32:39 crc kubenswrapper[5002]: I1014 09:32:39.219786 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.218720 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.219275 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.219358 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.219959 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.220012 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75" gracePeriod=600 Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.387157 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75" exitCode=0 Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.387228 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75"} Oct 14 09:33:09 crc kubenswrapper[5002]: I1014 09:33:09.387596 5002 scope.go:117] "RemoveContainer" containerID="170d324dca23bf9aef18b1bacf140b71ca3488ec3108ff99b5f1ab98a3e27fc2" Oct 14 09:33:10 crc kubenswrapper[5002]: I1014 09:33:10.402133 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805"} Oct 14 09:35:09 crc kubenswrapper[5002]: I1014 09:35:09.218454 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:35:09 crc kubenswrapper[5002]: I1014 09:35:09.219026 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:35:39 crc kubenswrapper[5002]: I1014 09:35:39.218531 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:35:39 crc kubenswrapper[5002]: I1014 09:35:39.219270 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.218493 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.219321 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.219431 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.220919 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.221090 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" gracePeriod=600 Oct 14 09:36:09 crc kubenswrapper[5002]: E1014 09:36:09.353429 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.402193 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" exitCode=0 Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.402240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805"} Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.402278 5002 scope.go:117] "RemoveContainer" containerID="739a23e59aecfad021fee49b2200f6e35bb79e7d94e04ee85bfa6a32a0557d75" Oct 14 09:36:09 crc kubenswrapper[5002]: I1014 09:36:09.403643 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:36:09 crc kubenswrapper[5002]: E1014 09:36:09.404606 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:36:19 crc kubenswrapper[5002]: I1014 09:36:19.721101 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:36:19 crc kubenswrapper[5002]: E1014 09:36:19.722582 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:36:34 crc kubenswrapper[5002]: I1014 09:36:34.721596 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:36:34 crc kubenswrapper[5002]: E1014 09:36:34.723025 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:36:46 crc kubenswrapper[5002]: I1014 09:36:46.720777 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:36:46 crc kubenswrapper[5002]: E1014 09:36:46.721621 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:36:57 crc kubenswrapper[5002]: I1014 09:36:57.726351 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:36:57 crc kubenswrapper[5002]: E1014 09:36:57.727987 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:37:09 crc kubenswrapper[5002]: I1014 09:37:09.722206 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:37:09 crc kubenswrapper[5002]: E1014 09:37:09.723102 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:37:24 crc kubenswrapper[5002]: I1014 09:37:24.720875 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:37:24 crc kubenswrapper[5002]: E1014 09:37:24.722208 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:37:39 crc kubenswrapper[5002]: I1014 09:37:39.721318 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:37:39 crc kubenswrapper[5002]: E1014 09:37:39.722441 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:37:50 crc kubenswrapper[5002]: I1014 09:37:50.721454 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:37:50 crc kubenswrapper[5002]: E1014 09:37:50.722606 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:38:04 crc kubenswrapper[5002]: I1014 09:38:04.721129 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:38:04 crc kubenswrapper[5002]: E1014 09:38:04.722531 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:38:19 crc kubenswrapper[5002]: I1014 09:38:19.721025 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:38:19 crc kubenswrapper[5002]: E1014 09:38:19.722256 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:38:33 crc kubenswrapper[5002]: I1014 09:38:33.721460 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:38:33 crc kubenswrapper[5002]: E1014 09:38:33.724276 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:38:46 crc kubenswrapper[5002]: I1014 09:38:46.720219 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:38:46 crc kubenswrapper[5002]: E1014 09:38:46.721059 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:38:58 crc kubenswrapper[5002]: I1014 09:38:58.722049 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:38:58 crc kubenswrapper[5002]: E1014 09:38:58.723243 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:39:13 crc kubenswrapper[5002]: I1014 09:39:13.721316 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:39:13 crc kubenswrapper[5002]: E1014 09:39:13.721972 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:39:25 crc kubenswrapper[5002]: I1014 09:39:25.720476 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:39:25 crc kubenswrapper[5002]: E1014 09:39:25.721153 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.447275 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451458 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="extract-content" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451485 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="extract-content" Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451507 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="extract-content" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451517 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="extract-content" Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451559 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451569 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451590 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="extract-utilities" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451599 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="extract-utilities" Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451617 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="extract-utilities" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451627 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="extract-utilities" Oct 14 09:39:33 crc kubenswrapper[5002]: E1014 09:39:33.451648 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451657 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451901 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c7d3c63-3c1c-4c17-9a48-91753cc29e4d" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.451938 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce05640d-53b3-4703-9b19-ea9b67b76aa5" containerName="registry-server" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.453698 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.471191 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.576405 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk677\" (UniqueName: \"kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.576478 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.576683 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.679366 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk677\" (UniqueName: \"kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.679430 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.679498 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.680203 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.680303 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.702094 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk677\" (UniqueName: \"kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677\") pod \"redhat-operators-9z2p9\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:33 crc kubenswrapper[5002]: I1014 09:39:33.788054 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:34 crc kubenswrapper[5002]: I1014 09:39:34.267702 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:34 crc kubenswrapper[5002]: I1014 09:39:34.637252 5002 generic.go:334] "Generic (PLEG): container finished" podID="06d6a355-db33-4fea-b665-839c32aa76f8" containerID="fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe" exitCode=0 Oct 14 09:39:34 crc kubenswrapper[5002]: I1014 09:39:34.637459 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerDied","Data":"fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe"} Oct 14 09:39:34 crc kubenswrapper[5002]: I1014 09:39:34.637482 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerStarted","Data":"be05b7e15c268b759d64dc5cc70f45d9db90442bb683cc8cb3599cf967956110"} Oct 14 09:39:34 crc kubenswrapper[5002]: I1014 09:39:34.639089 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:39:35 crc kubenswrapper[5002]: I1014 09:39:35.662990 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerStarted","Data":"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6"} Oct 14 09:39:37 crc kubenswrapper[5002]: I1014 09:39:37.728621 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:39:37 crc kubenswrapper[5002]: E1014 09:39:37.729572 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:39:38 crc kubenswrapper[5002]: I1014 09:39:38.692262 5002 generic.go:334] "Generic (PLEG): container finished" podID="06d6a355-db33-4fea-b665-839c32aa76f8" containerID="d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6" exitCode=0 Oct 14 09:39:38 crc kubenswrapper[5002]: I1014 09:39:38.692312 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerDied","Data":"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6"} Oct 14 09:39:39 crc kubenswrapper[5002]: I1014 09:39:39.704737 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerStarted","Data":"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27"} Oct 14 09:39:43 crc kubenswrapper[5002]: I1014 09:39:43.788217 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:43 crc kubenswrapper[5002]: I1014 09:39:43.788979 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:44 crc kubenswrapper[5002]: I1014 09:39:44.845811 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9z2p9" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="registry-server" probeResult="failure" output=< Oct 14 09:39:44 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:39:44 crc kubenswrapper[5002]: > Oct 14 09:39:49 crc kubenswrapper[5002]: I1014 09:39:49.721108 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:39:49 crc kubenswrapper[5002]: E1014 09:39:49.722065 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:39:53 crc kubenswrapper[5002]: I1014 09:39:53.874138 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:53 crc kubenswrapper[5002]: I1014 09:39:53.909581 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9z2p9" podStartSLOduration=16.312184287 podStartE2EDuration="20.909555515s" podCreationTimestamp="2025-10-14 09:39:33 +0000 UTC" firstStartedPulling="2025-10-14 09:39:34.638903974 +0000 UTC m=+6507.620143426" lastFinishedPulling="2025-10-14 09:39:39.236275192 +0000 UTC m=+6512.217514654" observedRunningTime="2025-10-14 09:39:39.743650823 +0000 UTC m=+6512.724890325" watchObservedRunningTime="2025-10-14 09:39:53.909555515 +0000 UTC m=+6526.890795007" Oct 14 09:39:53 crc kubenswrapper[5002]: I1014 09:39:53.952954 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:54 crc kubenswrapper[5002]: I1014 09:39:54.126050 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:55 crc kubenswrapper[5002]: I1014 09:39:55.890756 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9z2p9" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="registry-server" containerID="cri-o://4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27" gracePeriod=2 Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.420079 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.594052 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities\") pod \"06d6a355-db33-4fea-b665-839c32aa76f8\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.594254 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content\") pod \"06d6a355-db33-4fea-b665-839c32aa76f8\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.594283 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk677\" (UniqueName: \"kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677\") pod \"06d6a355-db33-4fea-b665-839c32aa76f8\" (UID: \"06d6a355-db33-4fea-b665-839c32aa76f8\") " Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.595498 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities" (OuterVolumeSpecName: "utilities") pod "06d6a355-db33-4fea-b665-839c32aa76f8" (UID: "06d6a355-db33-4fea-b665-839c32aa76f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.600330 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677" (OuterVolumeSpecName: "kube-api-access-zk677") pod "06d6a355-db33-4fea-b665-839c32aa76f8" (UID: "06d6a355-db33-4fea-b665-839c32aa76f8"). InnerVolumeSpecName "kube-api-access-zk677". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.697207 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.697260 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk677\" (UniqueName: \"kubernetes.io/projected/06d6a355-db33-4fea-b665-839c32aa76f8-kube-api-access-zk677\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.704603 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06d6a355-db33-4fea-b665-839c32aa76f8" (UID: "06d6a355-db33-4fea-b665-839c32aa76f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.799136 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d6a355-db33-4fea-b665-839c32aa76f8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.905349 5002 generic.go:334] "Generic (PLEG): container finished" podID="06d6a355-db33-4fea-b665-839c32aa76f8" containerID="4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27" exitCode=0 Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.905429 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerDied","Data":"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27"} Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.905483 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9z2p9" event={"ID":"06d6a355-db33-4fea-b665-839c32aa76f8","Type":"ContainerDied","Data":"be05b7e15c268b759d64dc5cc70f45d9db90442bb683cc8cb3599cf967956110"} Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.905521 5002 scope.go:117] "RemoveContainer" containerID="4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.905440 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9z2p9" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.938008 5002 scope.go:117] "RemoveContainer" containerID="d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6" Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.969251 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.985909 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9z2p9"] Oct 14 09:39:56 crc kubenswrapper[5002]: I1014 09:39:56.992197 5002 scope.go:117] "RemoveContainer" containerID="fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.046887 5002 scope.go:117] "RemoveContainer" containerID="4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27" Oct 14 09:39:57 crc kubenswrapper[5002]: E1014 09:39:57.047392 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27\": container with ID starting with 4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27 not found: ID does not exist" containerID="4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.047590 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27"} err="failed to get container status \"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27\": rpc error: code = NotFound desc = could not find container \"4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27\": container with ID starting with 4d53b404beee86e41b6b9c9f127b38e2e01e5678a630b42c713bb1c79e6e5d27 not found: ID does not exist" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.047744 5002 scope.go:117] "RemoveContainer" containerID="d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6" Oct 14 09:39:57 crc kubenswrapper[5002]: E1014 09:39:57.048469 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6\": container with ID starting with d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6 not found: ID does not exist" containerID="d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.048658 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6"} err="failed to get container status \"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6\": rpc error: code = NotFound desc = could not find container \"d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6\": container with ID starting with d5072faaea306a551d9bacc2c33267d7895e89de8636d8211273a666adae26e6 not found: ID does not exist" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.048797 5002 scope.go:117] "RemoveContainer" containerID="fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe" Oct 14 09:39:57 crc kubenswrapper[5002]: E1014 09:39:57.049370 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe\": container with ID starting with fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe not found: ID does not exist" containerID="fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.049455 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe"} err="failed to get container status \"fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe\": rpc error: code = NotFound desc = could not find container \"fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe\": container with ID starting with fca572aac01e7b612e2d3e1baa37ce44efbde761ffe1ad0c553179bbefb203fe not found: ID does not exist" Oct 14 09:39:57 crc kubenswrapper[5002]: I1014 09:39:57.736729 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" path="/var/lib/kubelet/pods/06d6a355-db33-4fea-b665-839c32aa76f8/volumes" Oct 14 09:40:01 crc kubenswrapper[5002]: I1014 09:40:01.721700 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:40:01 crc kubenswrapper[5002]: E1014 09:40:01.723696 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:40:13 crc kubenswrapper[5002]: I1014 09:40:13.721038 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:40:13 crc kubenswrapper[5002]: E1014 09:40:13.721775 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:40:27 crc kubenswrapper[5002]: I1014 09:40:27.736310 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:40:27 crc kubenswrapper[5002]: E1014 09:40:27.737147 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.320811 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:33 crc kubenswrapper[5002]: E1014 09:40:33.321818 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="extract-utilities" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.321833 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="extract-utilities" Oct 14 09:40:33 crc kubenswrapper[5002]: E1014 09:40:33.321874 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="extract-content" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.321883 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="extract-content" Oct 14 09:40:33 crc kubenswrapper[5002]: E1014 09:40:33.321900 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="registry-server" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.321909 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="registry-server" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.322177 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d6a355-db33-4fea-b665-839c32aa76f8" containerName="registry-server" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.323959 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.337713 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.498404 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6bqn\" (UniqueName: \"kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.498444 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.498545 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.600102 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6bqn\" (UniqueName: \"kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.600157 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.600222 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.600723 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.600764 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.621124 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6bqn\" (UniqueName: \"kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn\") pod \"community-operators-2z657\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:33 crc kubenswrapper[5002]: I1014 09:40:33.667081 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:34 crc kubenswrapper[5002]: I1014 09:40:34.237569 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:34 crc kubenswrapper[5002]: W1014 09:40:34.244312 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45cc1f67_373e_4231_a4d4_223c2a8e7b50.slice/crio-3f29a9165eec2e02208d1e4cd092f82b9849edcbc446d62ad31ab49b88fce123 WatchSource:0}: Error finding container 3f29a9165eec2e02208d1e4cd092f82b9849edcbc446d62ad31ab49b88fce123: Status 404 returned error can't find the container with id 3f29a9165eec2e02208d1e4cd092f82b9849edcbc446d62ad31ab49b88fce123 Oct 14 09:40:34 crc kubenswrapper[5002]: I1014 09:40:34.351215 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerStarted","Data":"3f29a9165eec2e02208d1e4cd092f82b9849edcbc446d62ad31ab49b88fce123"} Oct 14 09:40:35 crc kubenswrapper[5002]: I1014 09:40:35.364426 5002 generic.go:334] "Generic (PLEG): container finished" podID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerID="e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf" exitCode=0 Oct 14 09:40:35 crc kubenswrapper[5002]: I1014 09:40:35.364528 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerDied","Data":"e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf"} Oct 14 09:40:37 crc kubenswrapper[5002]: I1014 09:40:37.386734 5002 generic.go:334] "Generic (PLEG): container finished" podID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerID="0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721" exitCode=0 Oct 14 09:40:37 crc kubenswrapper[5002]: I1014 09:40:37.386799 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerDied","Data":"0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721"} Oct 14 09:40:38 crc kubenswrapper[5002]: I1014 09:40:38.398689 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerStarted","Data":"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c"} Oct 14 09:40:38 crc kubenswrapper[5002]: I1014 09:40:38.429942 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2z657" podStartSLOduration=2.9702592169999997 podStartE2EDuration="5.429909762s" podCreationTimestamp="2025-10-14 09:40:33 +0000 UTC" firstStartedPulling="2025-10-14 09:40:35.366819367 +0000 UTC m=+6568.348058819" lastFinishedPulling="2025-10-14 09:40:37.826469912 +0000 UTC m=+6570.807709364" observedRunningTime="2025-10-14 09:40:38.419208396 +0000 UTC m=+6571.400447848" watchObservedRunningTime="2025-10-14 09:40:38.429909762 +0000 UTC m=+6571.411149234" Oct 14 09:40:41 crc kubenswrapper[5002]: I1014 09:40:41.723159 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:40:41 crc kubenswrapper[5002]: E1014 09:40:41.725081 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:40:43 crc kubenswrapper[5002]: I1014 09:40:43.667873 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:43 crc kubenswrapper[5002]: I1014 09:40:43.668268 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:43 crc kubenswrapper[5002]: I1014 09:40:43.734262 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:44 crc kubenswrapper[5002]: I1014 09:40:44.547301 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:44 crc kubenswrapper[5002]: I1014 09:40:44.607482 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.392878 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.395070 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.420440 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.483624 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.484041 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqzqx\" (UniqueName: \"kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.484121 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.489975 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2z657" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="registry-server" containerID="cri-o://488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c" gracePeriod=2 Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.586497 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqzqx\" (UniqueName: \"kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.586586 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.586706 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.587237 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.587301 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.618754 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqzqx\" (UniqueName: \"kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx\") pod \"certified-operators-7z5f6\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.740325 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:46 crc kubenswrapper[5002]: I1014 09:40:46.968899 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.098738 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities\") pod \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.098790 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content\") pod \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.098964 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6bqn\" (UniqueName: \"kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn\") pod \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\" (UID: \"45cc1f67-373e-4231-a4d4-223c2a8e7b50\") " Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.104313 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn" (OuterVolumeSpecName: "kube-api-access-f6bqn") pod "45cc1f67-373e-4231-a4d4-223c2a8e7b50" (UID: "45cc1f67-373e-4231-a4d4-223c2a8e7b50"). InnerVolumeSpecName "kube-api-access-f6bqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.107784 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities" (OuterVolumeSpecName: "utilities") pod "45cc1f67-373e-4231-a4d4-223c2a8e7b50" (UID: "45cc1f67-373e-4231-a4d4-223c2a8e7b50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.201165 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6bqn\" (UniqueName: \"kubernetes.io/projected/45cc1f67-373e-4231-a4d4-223c2a8e7b50-kube-api-access-f6bqn\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.201194 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.283024 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.372548 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45cc1f67-373e-4231-a4d4-223c2a8e7b50" (UID: "45cc1f67-373e-4231-a4d4-223c2a8e7b50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.406314 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45cc1f67-373e-4231-a4d4-223c2a8e7b50-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.498342 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerID="d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c" exitCode=0 Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.498467 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerDied","Data":"d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c"} Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.498525 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerStarted","Data":"3dcb2e7582672b480d55734695c19afc417507ed6f58e88facf541e2842f147e"} Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.503550 5002 generic.go:334] "Generic (PLEG): container finished" podID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerID="488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c" exitCode=0 Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.503597 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerDied","Data":"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c"} Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.503615 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2z657" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.503633 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2z657" event={"ID":"45cc1f67-373e-4231-a4d4-223c2a8e7b50","Type":"ContainerDied","Data":"3f29a9165eec2e02208d1e4cd092f82b9849edcbc446d62ad31ab49b88fce123"} Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.503657 5002 scope.go:117] "RemoveContainer" containerID="488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.527310 5002 scope.go:117] "RemoveContainer" containerID="0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.542882 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.550003 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2z657"] Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.565175 5002 scope.go:117] "RemoveContainer" containerID="e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.582520 5002 scope.go:117] "RemoveContainer" containerID="488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c" Oct 14 09:40:47 crc kubenswrapper[5002]: E1014 09:40:47.582820 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c\": container with ID starting with 488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c not found: ID does not exist" containerID="488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.582874 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c"} err="failed to get container status \"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c\": rpc error: code = NotFound desc = could not find container \"488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c\": container with ID starting with 488e058d28c3db9de920e1d5aa1528f46d42a1b3e79612c2b4a3f4187343214c not found: ID does not exist" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.582898 5002 scope.go:117] "RemoveContainer" containerID="0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721" Oct 14 09:40:47 crc kubenswrapper[5002]: E1014 09:40:47.583243 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721\": container with ID starting with 0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721 not found: ID does not exist" containerID="0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.583265 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721"} err="failed to get container status \"0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721\": rpc error: code = NotFound desc = could not find container \"0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721\": container with ID starting with 0854e15ffb9e21947e9a2e0e3068e17777c7d9d50533765af321202e7d498721 not found: ID does not exist" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.583279 5002 scope.go:117] "RemoveContainer" containerID="e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf" Oct 14 09:40:47 crc kubenswrapper[5002]: E1014 09:40:47.583558 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf\": container with ID starting with e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf not found: ID does not exist" containerID="e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.583613 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf"} err="failed to get container status \"e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf\": rpc error: code = NotFound desc = could not find container \"e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf\": container with ID starting with e822f4c8a0a5387b0f3931769e2563864e3acf13a8792fede8fd94c5bb0171bf not found: ID does not exist" Oct 14 09:40:47 crc kubenswrapper[5002]: I1014 09:40:47.736746 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" path="/var/lib/kubelet/pods/45cc1f67-373e-4231-a4d4-223c2a8e7b50/volumes" Oct 14 09:40:48 crc kubenswrapper[5002]: I1014 09:40:48.515706 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerStarted","Data":"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd"} Oct 14 09:40:49 crc kubenswrapper[5002]: I1014 09:40:49.531322 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerID="605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd" exitCode=0 Oct 14 09:40:49 crc kubenswrapper[5002]: I1014 09:40:49.531431 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerDied","Data":"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd"} Oct 14 09:40:50 crc kubenswrapper[5002]: I1014 09:40:50.547012 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerStarted","Data":"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a"} Oct 14 09:40:50 crc kubenswrapper[5002]: I1014 09:40:50.568382 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7z5f6" podStartSLOduration=2.131929406 podStartE2EDuration="4.568354939s" podCreationTimestamp="2025-10-14 09:40:46 +0000 UTC" firstStartedPulling="2025-10-14 09:40:47.500092357 +0000 UTC m=+6580.481331819" lastFinishedPulling="2025-10-14 09:40:49.93651787 +0000 UTC m=+6582.917757352" observedRunningTime="2025-10-14 09:40:50.567060535 +0000 UTC m=+6583.548300047" watchObservedRunningTime="2025-10-14 09:40:50.568354939 +0000 UTC m=+6583.549594421" Oct 14 09:40:52 crc kubenswrapper[5002]: I1014 09:40:52.720406 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:40:52 crc kubenswrapper[5002]: E1014 09:40:52.722015 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:40:56 crc kubenswrapper[5002]: I1014 09:40:56.741398 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:56 crc kubenswrapper[5002]: I1014 09:40:56.742016 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:56 crc kubenswrapper[5002]: I1014 09:40:56.826772 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:57 crc kubenswrapper[5002]: I1014 09:40:57.674260 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:40:57 crc kubenswrapper[5002]: I1014 09:40:57.743285 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:40:59 crc kubenswrapper[5002]: I1014 09:40:59.664087 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7z5f6" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="registry-server" containerID="cri-o://36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a" gracePeriod=2 Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.337544 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.398399 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqzqx\" (UniqueName: \"kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx\") pod \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.398774 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content\") pod \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.405247 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities\") pod \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\" (UID: \"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e\") " Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.406056 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities" (OuterVolumeSpecName: "utilities") pod "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" (UID: "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.406520 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.410224 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx" (OuterVolumeSpecName: "kube-api-access-wqzqx") pod "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" (UID: "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e"). InnerVolumeSpecName "kube-api-access-wqzqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.481380 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" (UID: "7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.508941 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqzqx\" (UniqueName: \"kubernetes.io/projected/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-kube-api-access-wqzqx\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.508974 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.677154 5002 generic.go:334] "Generic (PLEG): container finished" podID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerID="36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a" exitCode=0 Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.677190 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerDied","Data":"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a"} Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.677226 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7z5f6" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.677250 5002 scope.go:117] "RemoveContainer" containerID="36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.677236 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7z5f6" event={"ID":"7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e","Type":"ContainerDied","Data":"3dcb2e7582672b480d55734695c19afc417507ed6f58e88facf541e2842f147e"} Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.710314 5002 scope.go:117] "RemoveContainer" containerID="605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.724723 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.739919 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7z5f6"] Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.743643 5002 scope.go:117] "RemoveContainer" containerID="d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.800095 5002 scope.go:117] "RemoveContainer" containerID="36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a" Oct 14 09:41:00 crc kubenswrapper[5002]: E1014 09:41:00.800551 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a\": container with ID starting with 36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a not found: ID does not exist" containerID="36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.800580 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a"} err="failed to get container status \"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a\": rpc error: code = NotFound desc = could not find container \"36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a\": container with ID starting with 36f54454a1b5467f9a052d4d23a1243462f188afc1f620fe11843cd2d0b0419a not found: ID does not exist" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.800600 5002 scope.go:117] "RemoveContainer" containerID="605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd" Oct 14 09:41:00 crc kubenswrapper[5002]: E1014 09:41:00.801056 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd\": container with ID starting with 605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd not found: ID does not exist" containerID="605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.801107 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd"} err="failed to get container status \"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd\": rpc error: code = NotFound desc = could not find container \"605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd\": container with ID starting with 605e07bf8b2438538961a418e9047ce523918bcc031416d6c9e7d376537681bd not found: ID does not exist" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.801140 5002 scope.go:117] "RemoveContainer" containerID="d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c" Oct 14 09:41:00 crc kubenswrapper[5002]: E1014 09:41:00.801640 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c\": container with ID starting with d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c not found: ID does not exist" containerID="d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c" Oct 14 09:41:00 crc kubenswrapper[5002]: I1014 09:41:00.801668 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c"} err="failed to get container status \"d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c\": rpc error: code = NotFound desc = could not find container \"d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c\": container with ID starting with d65e26217869474e9886e2b04b72e906728d78cb5150ad202a740fc0f7c99a7c not found: ID does not exist" Oct 14 09:41:01 crc kubenswrapper[5002]: I1014 09:41:01.731707 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" path="/var/lib/kubelet/pods/7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e/volumes" Oct 14 09:41:04 crc kubenswrapper[5002]: I1014 09:41:04.720886 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:41:04 crc kubenswrapper[5002]: E1014 09:41:04.722228 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:41:15 crc kubenswrapper[5002]: I1014 09:41:15.720794 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:41:16 crc kubenswrapper[5002]: I1014 09:41:16.891181 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781"} Oct 14 09:43:39 crc kubenswrapper[5002]: I1014 09:43:39.218379 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:43:39 crc kubenswrapper[5002]: I1014 09:43:39.219197 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:44:09 crc kubenswrapper[5002]: I1014 09:44:09.218993 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:44:09 crc kubenswrapper[5002]: I1014 09:44:09.219814 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:44:39 crc kubenswrapper[5002]: I1014 09:44:39.218815 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:44:39 crc kubenswrapper[5002]: I1014 09:44:39.219312 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:44:39 crc kubenswrapper[5002]: I1014 09:44:39.219361 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:44:39 crc kubenswrapper[5002]: I1014 09:44:39.220185 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:44:39 crc kubenswrapper[5002]: I1014 09:44:39.220243 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781" gracePeriod=600 Oct 14 09:44:40 crc kubenswrapper[5002]: I1014 09:44:40.153114 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781" exitCode=0 Oct 14 09:44:40 crc kubenswrapper[5002]: I1014 09:44:40.153924 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781"} Oct 14 09:44:40 crc kubenswrapper[5002]: I1014 09:44:40.153978 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5"} Oct 14 09:44:40 crc kubenswrapper[5002]: I1014 09:44:40.154011 5002 scope.go:117] "RemoveContainer" containerID="e852c4ae49b9a4c1747ada355da5bca4dd1d1bdd08c240848f21441ee5464805" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.222177 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7"] Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.223805 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.223887 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.223915 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.223927 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.223982 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.223997 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.224107 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.224121 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.224143 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.224197 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[5002]: E1014 09:45:00.224223 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.224235 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.224829 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="45cc1f67-373e-4231-a4d4-223c2a8e7b50" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.224926 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a1fe938-5f21-4ea7-8b48-9f10cc9fa22e" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.226567 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.229197 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.229361 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.234917 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7"] Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.352037 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.352102 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.352576 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jvb9\" (UniqueName: \"kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.454412 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jvb9\" (UniqueName: \"kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.454589 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.454648 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.456618 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.460285 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.475989 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jvb9\" (UniqueName: \"kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9\") pod \"collect-profiles-29340585-5s9s7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.554015 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:00 crc kubenswrapper[5002]: I1014 09:45:00.846168 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7"] Oct 14 09:45:01 crc kubenswrapper[5002]: I1014 09:45:01.388613 5002 generic.go:334] "Generic (PLEG): container finished" podID="01f99989-fa01-4fb3-b7e1-62759a90f3b7" containerID="465c6ee6b0505be46db95e2ca61bdebf75dc70f5ec4fea8796fb304095cb7a9e" exitCode=0 Oct 14 09:45:01 crc kubenswrapper[5002]: I1014 09:45:01.388675 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" event={"ID":"01f99989-fa01-4fb3-b7e1-62759a90f3b7","Type":"ContainerDied","Data":"465c6ee6b0505be46db95e2ca61bdebf75dc70f5ec4fea8796fb304095cb7a9e"} Oct 14 09:45:01 crc kubenswrapper[5002]: I1014 09:45:01.388706 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" event={"ID":"01f99989-fa01-4fb3-b7e1-62759a90f3b7","Type":"ContainerStarted","Data":"c8cda3b879629a8059be8fb92d33b2659a66a8ec5f80d19b2aa5fefb0961e752"} Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.822675 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.937045 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume\") pod \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.937144 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume\") pod \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.937192 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jvb9\" (UniqueName: \"kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9\") pod \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\" (UID: \"01f99989-fa01-4fb3-b7e1-62759a90f3b7\") " Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.938204 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume" (OuterVolumeSpecName: "config-volume") pod "01f99989-fa01-4fb3-b7e1-62759a90f3b7" (UID: "01f99989-fa01-4fb3-b7e1-62759a90f3b7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.958351 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "01f99989-fa01-4fb3-b7e1-62759a90f3b7" (UID: "01f99989-fa01-4fb3-b7e1-62759a90f3b7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:02 crc kubenswrapper[5002]: I1014 09:45:02.958403 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9" (OuterVolumeSpecName: "kube-api-access-4jvb9") pod "01f99989-fa01-4fb3-b7e1-62759a90f3b7" (UID: "01f99989-fa01-4fb3-b7e1-62759a90f3b7"). InnerVolumeSpecName "kube-api-access-4jvb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.039727 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/01f99989-fa01-4fb3-b7e1-62759a90f3b7-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.039771 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jvb9\" (UniqueName: \"kubernetes.io/projected/01f99989-fa01-4fb3-b7e1-62759a90f3b7-kube-api-access-4jvb9\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.039785 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/01f99989-fa01-4fb3-b7e1-62759a90f3b7-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.416603 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" event={"ID":"01f99989-fa01-4fb3-b7e1-62759a90f3b7","Type":"ContainerDied","Data":"c8cda3b879629a8059be8fb92d33b2659a66a8ec5f80d19b2aa5fefb0961e752"} Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.416645 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8cda3b879629a8059be8fb92d33b2659a66a8ec5f80d19b2aa5fefb0961e752" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.416666 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-5s9s7" Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.920781 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd"] Oct 14 09:45:03 crc kubenswrapper[5002]: I1014 09:45:03.930338 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-74twd"] Oct 14 09:45:05 crc kubenswrapper[5002]: I1014 09:45:05.736413 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d99e72de-1dbc-4ca2-8fd7-0f04e35d6576" path="/var/lib/kubelet/pods/d99e72de-1dbc-4ca2-8fd7-0f04e35d6576/volumes" Oct 14 09:45:30 crc kubenswrapper[5002]: I1014 09:45:30.919782 5002 scope.go:117] "RemoveContainer" containerID="d59821667e674e4349807a53c5eb7f3ebb2b80531840615c47dac5ad11573b84" Oct 14 09:45:32 crc kubenswrapper[5002]: I1014 09:45:32.767050 5002 generic.go:334] "Generic (PLEG): container finished" podID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" containerID="45f95251e4113681c6719ac32efb6accda6c204884b6beaa6ec54a9d93278500" exitCode=0 Oct 14 09:45:32 crc kubenswrapper[5002]: I1014 09:45:32.767166 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"ec2597cf-678d-44a5-a3e3-41be976a3dc6","Type":"ContainerDied","Data":"45f95251e4113681c6719ac32efb6accda6c204884b6beaa6ec54a9d93278500"} Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.340071 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.442263 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 14 09:45:34 crc kubenswrapper[5002]: E1014 09:45:34.443181 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01f99989-fa01-4fb3-b7e1-62759a90f3b7" containerName="collect-profiles" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.443215 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="01f99989-fa01-4fb3-b7e1-62759a90f3b7" containerName="collect-profiles" Oct 14 09:45:34 crc kubenswrapper[5002]: E1014 09:45:34.443263 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.443274 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.443599 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="01f99989-fa01-4fb3-b7e1-62759a90f3b7" containerName="collect-profiles" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.443649 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec2597cf-678d-44a5-a3e3-41be976a3dc6" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.445294 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.447509 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.447804 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.458300 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.490701 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwjb8\" (UniqueName: \"kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.491486 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.491945 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.492300 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.492563 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.492834 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.493076 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.493292 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.492684 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data" (OuterVolumeSpecName: "config-data") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.493748 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.494011 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\" (UID: \"ec2597cf-678d-44a5-a3e3-41be976a3dc6\") " Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.494311 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.494711 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.495183 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.495484 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.495990 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.496398 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.501696 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.516120 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph" (OuterVolumeSpecName: "ceph") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.516169 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.516245 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8" (OuterVolumeSpecName: "kube-api-access-zwjb8") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "kube-api-access-zwjb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.522287 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.523073 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.527685 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.541770 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ec2597cf-678d-44a5-a3e3-41be976a3dc6" (UID: "ec2597cf-678d-44a5-a3e3-41be976a3dc6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.597935 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598155 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598195 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598264 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598308 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598338 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598384 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598401 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598426 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598468 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clqhs\" (UniqueName: \"kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598522 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598535 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598543 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598552 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598561 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwjb8\" (UniqueName: \"kubernetes.io/projected/ec2597cf-678d-44a5-a3e3-41be976a3dc6-kube-api-access-zwjb8\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598571 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec2597cf-678d-44a5-a3e3-41be976a3dc6-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.598612 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec2597cf-678d-44a5-a3e3-41be976a3dc6-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.599727 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.600149 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.603198 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.642427 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700159 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700222 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700271 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700288 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700313 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.700330 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clqhs\" (UniqueName: \"kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.701004 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.701416 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.704565 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.705419 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.709239 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.715842 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clqhs\" (UniqueName: \"kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.765158 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.796073 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"ec2597cf-678d-44a5-a3e3-41be976a3dc6","Type":"ContainerDied","Data":"06ffbfddb91b06b31050e857f922d727725c302654a26a31cab154841fb669cb"} Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.796141 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Oct 14 09:45:34 crc kubenswrapper[5002]: I1014 09:45:34.796143 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06ffbfddb91b06b31050e857f922d727725c302654a26a31cab154841fb669cb" Oct 14 09:45:35 crc kubenswrapper[5002]: I1014 09:45:35.345181 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Oct 14 09:45:35 crc kubenswrapper[5002]: I1014 09:45:35.806835 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"08748a2d-35ba-47e7-a01a-51f71b61d041","Type":"ContainerStarted","Data":"6036b2ecb43ddc41394f82c58587bc73b35326c742b37bb20790bfab6d7fbe9e"} Oct 14 09:45:36 crc kubenswrapper[5002]: I1014 09:45:36.819127 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"08748a2d-35ba-47e7-a01a-51f71b61d041","Type":"ContainerStarted","Data":"11f2fd9d4cca6ae70fb0df3b6bc284f26fa4ed715d701b9445803d42a52a349a"} Oct 14 09:45:36 crc kubenswrapper[5002]: I1014 09:45:36.839720 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=2.839699881 podStartE2EDuration="2.839699881s" podCreationTimestamp="2025-10-14 09:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:45:36.831924434 +0000 UTC m=+6869.813163896" watchObservedRunningTime="2025-10-14 09:45:36.839699881 +0000 UTC m=+6869.820939333" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.107156 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.112117 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.117972 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.231131 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwmsv\" (UniqueName: \"kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.231188 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.231425 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.333507 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.333703 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwmsv\" (UniqueName: \"kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.333759 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.334095 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.334145 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.365585 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwmsv\" (UniqueName: \"kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv\") pod \"redhat-marketplace-5z6wd\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.441875 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:45:57 crc kubenswrapper[5002]: I1014 09:45:57.965233 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:45:58 crc kubenswrapper[5002]: I1014 09:45:58.074487 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerStarted","Data":"7be2375e6381ff44b5e7f9d92bd8c82b9189f53d2c838cc7755bc654a74e15e4"} Oct 14 09:45:59 crc kubenswrapper[5002]: I1014 09:45:59.096092 5002 generic.go:334] "Generic (PLEG): container finished" podID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerID="8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008" exitCode=0 Oct 14 09:45:59 crc kubenswrapper[5002]: I1014 09:45:59.096222 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerDied","Data":"8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008"} Oct 14 09:45:59 crc kubenswrapper[5002]: I1014 09:45:59.099801 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:46:01 crc kubenswrapper[5002]: I1014 09:46:01.129241 5002 generic.go:334] "Generic (PLEG): container finished" podID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerID="9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06" exitCode=0 Oct 14 09:46:01 crc kubenswrapper[5002]: I1014 09:46:01.129320 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerDied","Data":"9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06"} Oct 14 09:46:02 crc kubenswrapper[5002]: I1014 09:46:02.141696 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerStarted","Data":"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372"} Oct 14 09:46:02 crc kubenswrapper[5002]: I1014 09:46:02.176094 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5z6wd" podStartSLOduration=2.439976546 podStartE2EDuration="5.176066974s" podCreationTimestamp="2025-10-14 09:45:57 +0000 UTC" firstStartedPulling="2025-10-14 09:45:59.099305714 +0000 UTC m=+6892.080545206" lastFinishedPulling="2025-10-14 09:46:01.835396182 +0000 UTC m=+6894.816635634" observedRunningTime="2025-10-14 09:46:02.169245601 +0000 UTC m=+6895.150485053" watchObservedRunningTime="2025-10-14 09:46:02.176066974 +0000 UTC m=+6895.157306446" Oct 14 09:46:07 crc kubenswrapper[5002]: I1014 09:46:07.442276 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:07 crc kubenswrapper[5002]: I1014 09:46:07.442773 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:07 crc kubenswrapper[5002]: I1014 09:46:07.500293 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:08 crc kubenswrapper[5002]: I1014 09:46:08.264231 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:08 crc kubenswrapper[5002]: I1014 09:46:08.329600 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.228791 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5z6wd" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="registry-server" containerID="cri-o://28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372" gracePeriod=2 Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.727468 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.837739 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities\") pod \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.837928 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwmsv\" (UniqueName: \"kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv\") pod \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.838021 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content\") pod \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\" (UID: \"a32f964a-0d69-4d20-a55b-b20cf9c0a026\") " Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.839415 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities" (OuterVolumeSpecName: "utilities") pod "a32f964a-0d69-4d20-a55b-b20cf9c0a026" (UID: "a32f964a-0d69-4d20-a55b-b20cf9c0a026"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.842004 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.843928 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv" (OuterVolumeSpecName: "kube-api-access-kwmsv") pod "a32f964a-0d69-4d20-a55b-b20cf9c0a026" (UID: "a32f964a-0d69-4d20-a55b-b20cf9c0a026"). InnerVolumeSpecName "kube-api-access-kwmsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.856003 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a32f964a-0d69-4d20-a55b-b20cf9c0a026" (UID: "a32f964a-0d69-4d20-a55b-b20cf9c0a026"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.943424 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwmsv\" (UniqueName: \"kubernetes.io/projected/a32f964a-0d69-4d20-a55b-b20cf9c0a026-kube-api-access-kwmsv\") on node \"crc\" DevicePath \"\"" Oct 14 09:46:10 crc kubenswrapper[5002]: I1014 09:46:10.943460 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a32f964a-0d69-4d20-a55b-b20cf9c0a026-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.240230 5002 generic.go:334] "Generic (PLEG): container finished" podID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerID="28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372" exitCode=0 Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.240294 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerDied","Data":"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372"} Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.240567 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5z6wd" event={"ID":"a32f964a-0d69-4d20-a55b-b20cf9c0a026","Type":"ContainerDied","Data":"7be2375e6381ff44b5e7f9d92bd8c82b9189f53d2c838cc7755bc654a74e15e4"} Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.240614 5002 scope.go:117] "RemoveContainer" containerID="28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.240339 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5z6wd" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.279851 5002 scope.go:117] "RemoveContainer" containerID="9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.280819 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.290386 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5z6wd"] Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.308988 5002 scope.go:117] "RemoveContainer" containerID="8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.372957 5002 scope.go:117] "RemoveContainer" containerID="28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372" Oct 14 09:46:11 crc kubenswrapper[5002]: E1014 09:46:11.373453 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372\": container with ID starting with 28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372 not found: ID does not exist" containerID="28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.373504 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372"} err="failed to get container status \"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372\": rpc error: code = NotFound desc = could not find container \"28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372\": container with ID starting with 28cd00ed4bf31d1fada62a8d7650b39174448cea507a2c3e702d49cceff37372 not found: ID does not exist" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.373536 5002 scope.go:117] "RemoveContainer" containerID="9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06" Oct 14 09:46:11 crc kubenswrapper[5002]: E1014 09:46:11.374211 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06\": container with ID starting with 9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06 not found: ID does not exist" containerID="9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.374246 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06"} err="failed to get container status \"9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06\": rpc error: code = NotFound desc = could not find container \"9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06\": container with ID starting with 9408ca95d00105d2bc93e33e6e7b828265ed785120f6043b0079590f317f0a06 not found: ID does not exist" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.374272 5002 scope.go:117] "RemoveContainer" containerID="8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008" Oct 14 09:46:11 crc kubenswrapper[5002]: E1014 09:46:11.374718 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008\": container with ID starting with 8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008 not found: ID does not exist" containerID="8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.375007 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008"} err="failed to get container status \"8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008\": rpc error: code = NotFound desc = could not find container \"8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008\": container with ID starting with 8b300cfb3ed9b3b6c406ce8ff467501c0d47ba72e74456da0305b31e38626008 not found: ID does not exist" Oct 14 09:46:11 crc kubenswrapper[5002]: I1014 09:46:11.742532 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" path="/var/lib/kubelet/pods/a32f964a-0d69-4d20-a55b-b20cf9c0a026/volumes" Oct 14 09:46:39 crc kubenswrapper[5002]: I1014 09:46:39.218204 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:46:39 crc kubenswrapper[5002]: I1014 09:46:39.218805 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:47:09 crc kubenswrapper[5002]: I1014 09:47:09.218141 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:47:09 crc kubenswrapper[5002]: I1014 09:47:09.218780 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:47:39 crc kubenswrapper[5002]: I1014 09:47:39.218626 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:47:39 crc kubenswrapper[5002]: I1014 09:47:39.219380 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:47:39 crc kubenswrapper[5002]: I1014 09:47:39.219467 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:47:39 crc kubenswrapper[5002]: I1014 09:47:39.220648 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:47:39 crc kubenswrapper[5002]: I1014 09:47:39.220747 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" gracePeriod=600 Oct 14 09:47:39 crc kubenswrapper[5002]: E1014 09:47:39.355726 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:47:40 crc kubenswrapper[5002]: I1014 09:47:40.367913 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" exitCode=0 Oct 14 09:47:40 crc kubenswrapper[5002]: I1014 09:47:40.368012 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5"} Oct 14 09:47:40 crc kubenswrapper[5002]: I1014 09:47:40.368339 5002 scope.go:117] "RemoveContainer" containerID="254ef9d2db386cdfc24b758a860d0427fb0ea3df4f0addc30555eae40d2b2781" Oct 14 09:47:40 crc kubenswrapper[5002]: I1014 09:47:40.369179 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:47:40 crc kubenswrapper[5002]: E1014 09:47:40.369691 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:47:52 crc kubenswrapper[5002]: I1014 09:47:52.721916 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:47:52 crc kubenswrapper[5002]: E1014 09:47:52.722979 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:48:03 crc kubenswrapper[5002]: I1014 09:48:03.734314 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:48:03 crc kubenswrapper[5002]: E1014 09:48:03.735248 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:48:18 crc kubenswrapper[5002]: I1014 09:48:18.721367 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:48:18 crc kubenswrapper[5002]: E1014 09:48:18.722423 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:48:32 crc kubenswrapper[5002]: I1014 09:48:32.721274 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:48:32 crc kubenswrapper[5002]: E1014 09:48:32.722336 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:48:47 crc kubenswrapper[5002]: I1014 09:48:47.738758 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:48:47 crc kubenswrapper[5002]: E1014 09:48:47.740069 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:48:58 crc kubenswrapper[5002]: I1014 09:48:58.725043 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:48:58 crc kubenswrapper[5002]: E1014 09:48:58.726303 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:49:10 crc kubenswrapper[5002]: I1014 09:49:10.720752 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:49:10 crc kubenswrapper[5002]: E1014 09:49:10.722119 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:49:23 crc kubenswrapper[5002]: I1014 09:49:23.721061 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:49:23 crc kubenswrapper[5002]: E1014 09:49:23.722532 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:49:37 crc kubenswrapper[5002]: I1014 09:49:37.729059 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:49:37 crc kubenswrapper[5002]: E1014 09:49:37.730019 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:49:48 crc kubenswrapper[5002]: I1014 09:49:48.720254 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:49:48 crc kubenswrapper[5002]: E1014 09:49:48.721290 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:49:59 crc kubenswrapper[5002]: I1014 09:49:59.721501 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:49:59 crc kubenswrapper[5002]: E1014 09:49:59.722793 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:50:06 crc kubenswrapper[5002]: I1014 09:50:06.146034 5002 generic.go:334] "Generic (PLEG): container finished" podID="08748a2d-35ba-47e7-a01a-51f71b61d041" containerID="11f2fd9d4cca6ae70fb0df3b6bc284f26fa4ed715d701b9445803d42a52a349a" exitCode=0 Oct 14 09:50:06 crc kubenswrapper[5002]: I1014 09:50:06.146133 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"08748a2d-35ba-47e7-a01a-51f71b61d041","Type":"ContainerDied","Data":"11f2fd9d4cca6ae70fb0df3b6bc284f26fa4ed715d701b9445803d42a52a349a"} Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.752131 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.805076 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.805478 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.805589 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.805771 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.806086 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.807980 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.806350 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.808138 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clqhs\" (UniqueName: \"kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.808168 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.808262 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.808307 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data\") pod \"08748a2d-35ba-47e7-a01a-51f71b61d041\" (UID: \"08748a2d-35ba-47e7-a01a-51f71b61d041\") " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.809310 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.810076 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data" (OuterVolumeSpecName: "config-data") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.812051 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph" (OuterVolumeSpecName: "ceph") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.812719 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.819093 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs" (OuterVolumeSpecName: "kube-api-access-clqhs") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "kube-api-access-clqhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.819551 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.841911 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.852466 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.857218 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.867876 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "08748a2d-35ba-47e7-a01a-51f71b61d041" (UID: "08748a2d-35ba-47e7-a01a-51f71b61d041"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911721 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/08748a2d-35ba-47e7-a01a-51f71b61d041-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911754 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911768 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clqhs\" (UniqueName: \"kubernetes.io/projected/08748a2d-35ba-47e7-a01a-51f71b61d041-kube-api-access-clqhs\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911780 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911790 5002 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911805 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911858 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911870 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/08748a2d-35ba-47e7-a01a-51f71b61d041-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.911881 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/08748a2d-35ba-47e7-a01a-51f71b61d041-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:07 crc kubenswrapper[5002]: I1014 09:50:07.932754 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 14 09:50:08 crc kubenswrapper[5002]: I1014 09:50:08.013077 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:08 crc kubenswrapper[5002]: I1014 09:50:08.172865 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"08748a2d-35ba-47e7-a01a-51f71b61d041","Type":"ContainerDied","Data":"6036b2ecb43ddc41394f82c58587bc73b35326c742b37bb20790bfab6d7fbe9e"} Oct 14 09:50:08 crc kubenswrapper[5002]: I1014 09:50:08.172909 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Oct 14 09:50:08 crc kubenswrapper[5002]: I1014 09:50:08.172923 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6036b2ecb43ddc41394f82c58587bc73b35326c742b37bb20790bfab6d7fbe9e" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.400862 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 09:50:12 crc kubenswrapper[5002]: E1014 09:50:12.401909 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="registry-server" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.401928 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="registry-server" Oct 14 09:50:12 crc kubenswrapper[5002]: E1014 09:50:12.401948 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08748a2d-35ba-47e7-a01a-51f71b61d041" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.401956 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="08748a2d-35ba-47e7-a01a-51f71b61d041" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:50:12 crc kubenswrapper[5002]: E1014 09:50:12.401972 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="extract-content" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.401982 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="extract-content" Oct 14 09:50:12 crc kubenswrapper[5002]: E1014 09:50:12.402003 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="extract-utilities" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.402011 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="extract-utilities" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.402281 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="08748a2d-35ba-47e7-a01a-51f71b61d041" containerName="tempest-tests-tempest-tests-runner" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.402295 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32f964a-0d69-4d20-a55b-b20cf9c0a026" containerName="registry-server" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.403078 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.405819 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7gjvf" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.430970 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.525378 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.525486 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwk8g\" (UniqueName: \"kubernetes.io/projected/c8166676-1bb1-4eec-9d75-28ccb7357ec3-kube-api-access-gwk8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.626999 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwk8g\" (UniqueName: \"kubernetes.io/projected/c8166676-1bb1-4eec-9d75-28ccb7357ec3-kube-api-access-gwk8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.627149 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.627611 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.661382 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwk8g\" (UniqueName: \"kubernetes.io/projected/c8166676-1bb1-4eec-9d75-28ccb7357ec3-kube-api-access-gwk8g\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.672079 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c8166676-1bb1-4eec-9d75-28ccb7357ec3\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:12 crc kubenswrapper[5002]: I1014 09:50:12.727708 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 09:50:13 crc kubenswrapper[5002]: I1014 09:50:13.199373 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 09:50:13 crc kubenswrapper[5002]: W1014 09:50:13.205112 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8166676_1bb1_4eec_9d75_28ccb7357ec3.slice/crio-9739ffe3d2c55764d86fb1d0db92f7d4176ba2c11f0a0af76768e21ad9b7b368 WatchSource:0}: Error finding container 9739ffe3d2c55764d86fb1d0db92f7d4176ba2c11f0a0af76768e21ad9b7b368: Status 404 returned error can't find the container with id 9739ffe3d2c55764d86fb1d0db92f7d4176ba2c11f0a0af76768e21ad9b7b368 Oct 14 09:50:13 crc kubenswrapper[5002]: I1014 09:50:13.238683 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c8166676-1bb1-4eec-9d75-28ccb7357ec3","Type":"ContainerStarted","Data":"9739ffe3d2c55764d86fb1d0db92f7d4176ba2c11f0a0af76768e21ad9b7b368"} Oct 14 09:50:14 crc kubenswrapper[5002]: I1014 09:50:14.721564 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:50:14 crc kubenswrapper[5002]: E1014 09:50:14.722371 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:50:15 crc kubenswrapper[5002]: I1014 09:50:15.268240 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c8166676-1bb1-4eec-9d75-28ccb7357ec3","Type":"ContainerStarted","Data":"06616e0bcfce473970ff6e207290d73168b3d6e1ac5290fd4bbb43a67b333748"} Oct 14 09:50:15 crc kubenswrapper[5002]: I1014 09:50:15.304916 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.298974401 podStartE2EDuration="3.304879558s" podCreationTimestamp="2025-10-14 09:50:12 +0000 UTC" firstStartedPulling="2025-10-14 09:50:13.207926141 +0000 UTC m=+7146.189165613" lastFinishedPulling="2025-10-14 09:50:14.213831288 +0000 UTC m=+7147.195070770" observedRunningTime="2025-10-14 09:50:15.29135927 +0000 UTC m=+7148.272598792" watchObservedRunningTime="2025-10-14 09:50:15.304879558 +0000 UTC m=+7148.286119060" Oct 14 09:50:25 crc kubenswrapper[5002]: I1014 09:50:25.720851 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:50:25 crc kubenswrapper[5002]: E1014 09:50:25.722075 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.764625 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.767417 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.776925 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.776987 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.776925 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.777246 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.777521 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.777690 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.872724 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.872792 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.872948 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.872991 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873043 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873185 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873322 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873382 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873562 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdlqp\" (UniqueName: \"kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873688 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873754 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.873897 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.974763 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.974915 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdlqp\" (UniqueName: \"kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.974994 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975031 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975082 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975151 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975187 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975249 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975307 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975359 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975417 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975487 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.975714 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.977402 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.977400 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.977554 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.977713 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.978144 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.978570 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.981515 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.982058 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.992329 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:35 crc kubenswrapper[5002]: I1014 09:50:35.992727 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:36 crc kubenswrapper[5002]: I1014 09:50:36.007413 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdlqp\" (UniqueName: \"kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:36 crc kubenswrapper[5002]: I1014 09:50:36.013198 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:36 crc kubenswrapper[5002]: I1014 09:50:36.100928 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:50:36 crc kubenswrapper[5002]: I1014 09:50:36.714318 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Oct 14 09:50:36 crc kubenswrapper[5002]: W1014 09:50:36.732429 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c2247d7_13c4_4c8d_ab13_3f2fbd020a1e.slice/crio-61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371 WatchSource:0}: Error finding container 61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371: Status 404 returned error can't find the container with id 61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371 Oct 14 09:50:37 crc kubenswrapper[5002]: I1014 09:50:37.543765 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e","Type":"ContainerStarted","Data":"61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371"} Oct 14 09:50:39 crc kubenswrapper[5002]: I1014 09:50:39.721275 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:50:39 crc kubenswrapper[5002]: E1014 09:50:39.723763 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:50:50 crc kubenswrapper[5002]: I1014 09:50:50.721022 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:50:50 crc kubenswrapper[5002]: E1014 09:50:50.722487 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:51:01 crc kubenswrapper[5002]: I1014 09:51:01.721178 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:51:01 crc kubenswrapper[5002]: E1014 09:51:01.722323 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.769446 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.772510 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.783156 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.927548 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvrdx\" (UniqueName: \"kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.927599 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:11 crc kubenswrapper[5002]: I1014 09:51:11.927657 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.029177 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvrdx\" (UniqueName: \"kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.029472 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.029503 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.030322 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.030406 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.059653 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvrdx\" (UniqueName: \"kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx\") pod \"community-operators-nxqwk\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:12 crc kubenswrapper[5002]: I1014 09:51:12.098286 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:14 crc kubenswrapper[5002]: I1014 09:51:14.722786 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:51:14 crc kubenswrapper[5002]: E1014 09:51:14.723542 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:51:15 crc kubenswrapper[5002]: E1014 09:51:15.558009 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Oct 14 09:51:15 crc kubenswrapper[5002]: E1014 09:51:15.559296 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdlqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 09:51:15 crc kubenswrapper[5002]: E1014 09:51:15.560612 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" Oct 14 09:51:15 crc kubenswrapper[5002]: I1014 09:51:15.756031 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:15 crc kubenswrapper[5002]: W1014 09:51:15.756398 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d20ab47_f19c_4bae_8428_10f2bb9857bf.slice/crio-fa33042310804bbf152a1b7da1cf8635b27bf67896ef83da812cda80148c7f82 WatchSource:0}: Error finding container fa33042310804bbf152a1b7da1cf8635b27bf67896ef83da812cda80148c7f82: Status 404 returned error can't find the container with id fa33042310804bbf152a1b7da1cf8635b27bf67896ef83da812cda80148c7f82 Oct 14 09:51:15 crc kubenswrapper[5002]: I1014 09:51:15.968302 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerStarted","Data":"fa33042310804bbf152a1b7da1cf8635b27bf67896ef83da812cda80148c7f82"} Oct 14 09:51:15 crc kubenswrapper[5002]: E1014 09:51:15.970419 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" Oct 14 09:51:16 crc kubenswrapper[5002]: I1014 09:51:16.979796 5002 generic.go:334] "Generic (PLEG): container finished" podID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerID="6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1" exitCode=0 Oct 14 09:51:16 crc kubenswrapper[5002]: I1014 09:51:16.979885 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerDied","Data":"6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1"} Oct 14 09:51:16 crc kubenswrapper[5002]: I1014 09:51:16.983189 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:51:17 crc kubenswrapper[5002]: I1014 09:51:17.993666 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerStarted","Data":"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698"} Oct 14 09:51:19 crc kubenswrapper[5002]: I1014 09:51:19.009328 5002 generic.go:334] "Generic (PLEG): container finished" podID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerID="379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698" exitCode=0 Oct 14 09:51:19 crc kubenswrapper[5002]: I1014 09:51:19.009437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerDied","Data":"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698"} Oct 14 09:51:20 crc kubenswrapper[5002]: I1014 09:51:20.023798 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerStarted","Data":"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c"} Oct 14 09:51:20 crc kubenswrapper[5002]: I1014 09:51:20.070272 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nxqwk" podStartSLOduration=6.591961274 podStartE2EDuration="9.070237148s" podCreationTimestamp="2025-10-14 09:51:11 +0000 UTC" firstStartedPulling="2025-10-14 09:51:16.981825156 +0000 UTC m=+7209.963064648" lastFinishedPulling="2025-10-14 09:51:19.46010103 +0000 UTC m=+7212.441340522" observedRunningTime="2025-10-14 09:51:20.058063874 +0000 UTC m=+7213.039303346" watchObservedRunningTime="2025-10-14 09:51:20.070237148 +0000 UTC m=+7213.051476670" Oct 14 09:51:22 crc kubenswrapper[5002]: I1014 09:51:22.101086 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:22 crc kubenswrapper[5002]: I1014 09:51:22.101419 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:23 crc kubenswrapper[5002]: I1014 09:51:23.156305 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nxqwk" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="registry-server" probeResult="failure" output=< Oct 14 09:51:23 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:51:23 crc kubenswrapper[5002]: > Oct 14 09:51:29 crc kubenswrapper[5002]: I1014 09:51:29.721035 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:51:29 crc kubenswrapper[5002]: E1014 09:51:29.721833 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:51:30 crc kubenswrapper[5002]: I1014 09:51:30.136618 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e","Type":"ContainerStarted","Data":"af2cc0a5c7cd355fd487ff13e96389816a6b2d3d219d3c270e8998e970c4e96a"} Oct 14 09:51:30 crc kubenswrapper[5002]: I1014 09:51:30.172675 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=3.599538959 podStartE2EDuration="56.172646713s" podCreationTimestamp="2025-10-14 09:50:34 +0000 UTC" firstStartedPulling="2025-10-14 09:50:36.735244896 +0000 UTC m=+7169.716484368" lastFinishedPulling="2025-10-14 09:51:29.30835267 +0000 UTC m=+7222.289592122" observedRunningTime="2025-10-14 09:51:30.163942822 +0000 UTC m=+7223.145182314" watchObservedRunningTime="2025-10-14 09:51:30.172646713 +0000 UTC m=+7223.153886205" Oct 14 09:51:32 crc kubenswrapper[5002]: I1014 09:51:32.166706 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:32 crc kubenswrapper[5002]: I1014 09:51:32.220113 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:32 crc kubenswrapper[5002]: I1014 09:51:32.425021 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:34 crc kubenswrapper[5002]: I1014 09:51:34.177792 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nxqwk" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="registry-server" containerID="cri-o://998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c" gracePeriod=2 Oct 14 09:51:34 crc kubenswrapper[5002]: I1014 09:51:34.803389 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.003430 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities\") pod \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.003770 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content\") pod \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.003930 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvrdx\" (UniqueName: \"kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx\") pod \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\" (UID: \"4d20ab47-f19c-4bae-8428-10f2bb9857bf\") " Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.009165 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities" (OuterVolumeSpecName: "utilities") pod "4d20ab47-f19c-4bae-8428-10f2bb9857bf" (UID: "4d20ab47-f19c-4bae-8428-10f2bb9857bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.010357 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx" (OuterVolumeSpecName: "kube-api-access-bvrdx") pod "4d20ab47-f19c-4bae-8428-10f2bb9857bf" (UID: "4d20ab47-f19c-4bae-8428-10f2bb9857bf"). InnerVolumeSpecName "kube-api-access-bvrdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.077109 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d20ab47-f19c-4bae-8428-10f2bb9857bf" (UID: "4d20ab47-f19c-4bae-8428-10f2bb9857bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.106564 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.106598 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d20ab47-f19c-4bae-8428-10f2bb9857bf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.106609 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvrdx\" (UniqueName: \"kubernetes.io/projected/4d20ab47-f19c-4bae-8428-10f2bb9857bf-kube-api-access-bvrdx\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.190378 5002 generic.go:334] "Generic (PLEG): container finished" podID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerID="998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c" exitCode=0 Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.190433 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerDied","Data":"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c"} Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.190442 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nxqwk" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.190484 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nxqwk" event={"ID":"4d20ab47-f19c-4bae-8428-10f2bb9857bf","Type":"ContainerDied","Data":"fa33042310804bbf152a1b7da1cf8635b27bf67896ef83da812cda80148c7f82"} Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.190516 5002 scope.go:117] "RemoveContainer" containerID="998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.239308 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.248679 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nxqwk"] Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.614424 5002 scope.go:117] "RemoveContainer" containerID="379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.668572 5002 scope.go:117] "RemoveContainer" containerID="6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.709391 5002 scope.go:117] "RemoveContainer" containerID="998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c" Oct 14 09:51:35 crc kubenswrapper[5002]: E1014 09:51:35.709878 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c\": container with ID starting with 998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c not found: ID does not exist" containerID="998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.709917 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c"} err="failed to get container status \"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c\": rpc error: code = NotFound desc = could not find container \"998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c\": container with ID starting with 998f7b02dd7c78e28156455d91c208ca5a68fb0b77feb57b28aeb4213beaba9c not found: ID does not exist" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.709943 5002 scope.go:117] "RemoveContainer" containerID="379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698" Oct 14 09:51:35 crc kubenswrapper[5002]: E1014 09:51:35.710187 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698\": container with ID starting with 379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698 not found: ID does not exist" containerID="379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.710214 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698"} err="failed to get container status \"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698\": rpc error: code = NotFound desc = could not find container \"379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698\": container with ID starting with 379b4f730cf220e6fba65295a733d49e38bffd98469a43e71c122e17c7443698 not found: ID does not exist" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.710233 5002 scope.go:117] "RemoveContainer" containerID="6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1" Oct 14 09:51:35 crc kubenswrapper[5002]: E1014 09:51:35.710504 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1\": container with ID starting with 6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1 not found: ID does not exist" containerID="6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.710524 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1"} err="failed to get container status \"6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1\": rpc error: code = NotFound desc = could not find container \"6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1\": container with ID starting with 6020948e78f6216cc50dd449397f874b601e32d5ade733dc29c78ed36b6c7ea1 not found: ID does not exist" Oct 14 09:51:35 crc kubenswrapper[5002]: I1014 09:51:35.743735 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" path="/var/lib/kubelet/pods/4d20ab47-f19c-4bae-8428-10f2bb9857bf/volumes" Oct 14 09:51:44 crc kubenswrapper[5002]: I1014 09:51:44.720793 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:51:44 crc kubenswrapper[5002]: E1014 09:51:44.721649 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:51:55 crc kubenswrapper[5002]: I1014 09:51:55.721440 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:51:55 crc kubenswrapper[5002]: E1014 09:51:55.722824 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:52:06 crc kubenswrapper[5002]: I1014 09:52:06.720999 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:52:06 crc kubenswrapper[5002]: E1014 09:52:06.722207 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:52:20 crc kubenswrapper[5002]: I1014 09:52:20.720487 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:52:20 crc kubenswrapper[5002]: E1014 09:52:20.721528 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:52:35 crc kubenswrapper[5002]: I1014 09:52:35.721116 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:52:35 crc kubenswrapper[5002]: E1014 09:52:35.722048 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:52:42 crc kubenswrapper[5002]: I1014 09:52:42.969038 5002 generic.go:334] "Generic (PLEG): container finished" podID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" containerID="af2cc0a5c7cd355fd487ff13e96389816a6b2d3d219d3c270e8998e970c4e96a" exitCode=0 Oct 14 09:52:42 crc kubenswrapper[5002]: I1014 09:52:42.969739 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e","Type":"ContainerDied","Data":"af2cc0a5c7cd355fd487ff13e96389816a6b2d3d219d3c270e8998e970c4e96a"} Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.578023 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652151 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 14 09:52:44 crc kubenswrapper[5002]: E1014 09:52:44.652524 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="extract-utilities" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652549 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="extract-utilities" Oct 14 09:52:44 crc kubenswrapper[5002]: E1014 09:52:44.652562 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" containerName="tobiko-tests-tobiko" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652567 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" containerName="tobiko-tests-tobiko" Oct 14 09:52:44 crc kubenswrapper[5002]: E1014 09:52:44.652596 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="extract-content" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652603 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="extract-content" Oct 14 09:52:44 crc kubenswrapper[5002]: E1014 09:52:44.652616 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="registry-server" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652622 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="registry-server" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652864 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" containerName="tobiko-tests-tobiko" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.652890 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d20ab47-f19c-4bae-8428-10f2bb9857bf" containerName="registry-server" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.653523 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.665589 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674286 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674355 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674429 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674470 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674561 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdlqp\" (UniqueName: \"kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674599 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674633 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674709 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674761 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674789 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674823 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.674903 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs\") pod \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\" (UID: \"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e\") " Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.676420 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.677960 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.681375 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp" (OuterVolumeSpecName: "kube-api-access-rdlqp") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "kube-api-access-rdlqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.696320 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.703186 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph" (OuterVolumeSpecName: "ceph") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.711584 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.718772 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.721495 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.731797 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.738556 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.740083 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.744903 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780079 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780180 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780222 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780264 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780452 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780547 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780584 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtmpx\" (UniqueName: \"kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780607 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780663 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780697 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780722 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780742 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780794 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdlqp\" (UniqueName: \"kubernetes.io/projected/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kube-api-access-rdlqp\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780805 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780815 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780825 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780848 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780858 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780867 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780876 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.780884 5002 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.808620 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882516 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882599 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882672 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882697 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882726 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtmpx\" (UniqueName: \"kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882754 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882801 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882865 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882898 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.882923 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.883070 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.885262 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.885392 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.885687 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.890049 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.892076 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.892325 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.892763 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.892770 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.892807 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.894639 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.899368 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtmpx\" (UniqueName: \"kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.992299 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e","Type":"ContainerDied","Data":"61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371"} Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.992342 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61c7fca27ab34ea8e45813c39faae5f245b9fe17548d5a1403f6d5ac3b4f4371" Oct 14 09:52:44 crc kubenswrapper[5002]: I1014 09:52:44.992398 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Oct 14 09:52:45 crc kubenswrapper[5002]: I1014 09:52:45.134547 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:52:45 crc kubenswrapper[5002]: I1014 09:52:45.881043 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Oct 14 09:52:46 crc kubenswrapper[5002]: I1014 09:52:46.005123 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d8f34f80-d17b-4cdb-9834-f2e96de161ae","Type":"ContainerStarted","Data":"649ff974f90c8484e71a3344d8431dd1070b0aa894d2fd590a4d914265301b8e"} Oct 14 09:52:46 crc kubenswrapper[5002]: I1014 09:52:46.273515 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e" (UID: "8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:52:46 crc kubenswrapper[5002]: I1014 09:52:46.318043 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:47 crc kubenswrapper[5002]: I1014 09:52:47.019924 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d8f34f80-d17b-4cdb-9834-f2e96de161ae","Type":"ContainerStarted","Data":"a474a6bdb78ec7f91378160c4185c17fa9b6a1cb94be523eaa89de1bd28cc9a0"} Oct 14 09:52:47 crc kubenswrapper[5002]: I1014 09:52:47.067180 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.06716162 podStartE2EDuration="3.06716162s" podCreationTimestamp="2025-10-14 09:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:47.058603692 +0000 UTC m=+7300.039843184" watchObservedRunningTime="2025-10-14 09:52:47.06716162 +0000 UTC m=+7300.048401082" Oct 14 09:52:49 crc kubenswrapper[5002]: I1014 09:52:49.721363 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:52:50 crc kubenswrapper[5002]: I1014 09:52:50.054916 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e"} Oct 14 09:54:17 crc kubenswrapper[5002]: I1014 09:54:17.989520 5002 generic.go:334] "Generic (PLEG): container finished" podID="d8f34f80-d17b-4cdb-9834-f2e96de161ae" containerID="a474a6bdb78ec7f91378160c4185c17fa9b6a1cb94be523eaa89de1bd28cc9a0" exitCode=0 Oct 14 09:54:17 crc kubenswrapper[5002]: I1014 09:54:17.989660 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d8f34f80-d17b-4cdb-9834-f2e96de161ae","Type":"ContainerDied","Data":"a474a6bdb78ec7f91378160c4185c17fa9b6a1cb94be523eaa89de1bd28cc9a0"} Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.629082 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717436 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtmpx\" (UniqueName: \"kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717500 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717533 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717585 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717606 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717646 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717668 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717707 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717743 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717771 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717824 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.717856 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary\") pod \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\" (UID: \"d8f34f80-d17b-4cdb-9834-f2e96de161ae\") " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.719030 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.724007 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx" (OuterVolumeSpecName: "kube-api-access-gtmpx") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "kube-api-access-gtmpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.724446 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.745084 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph" (OuterVolumeSpecName: "ceph") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.750781 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.752865 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.756989 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.770330 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.787736 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.806751 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.809899 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821066 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtmpx\" (UniqueName: \"kubernetes.io/projected/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kube-api-access-gtmpx\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821123 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821154 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821196 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821212 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821226 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821240 5002 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-kubeconfig\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821252 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821312 5002 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/d8f34f80-d17b-4cdb-9834-f2e96de161ae-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821325 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d8f34f80-d17b-4cdb-9834-f2e96de161ae-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.821337 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.842037 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 14 09:54:19 crc kubenswrapper[5002]: I1014 09:54:19.925583 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:20 crc kubenswrapper[5002]: I1014 09:54:20.030988 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"d8f34f80-d17b-4cdb-9834-f2e96de161ae","Type":"ContainerDied","Data":"649ff974f90c8484e71a3344d8431dd1070b0aa894d2fd590a4d914265301b8e"} Oct 14 09:54:20 crc kubenswrapper[5002]: I1014 09:54:20.031026 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="649ff974f90c8484e71a3344d8431dd1070b0aa894d2fd590a4d914265301b8e" Oct 14 09:54:20 crc kubenswrapper[5002]: I1014 09:54:20.031044 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Oct 14 09:54:21 crc kubenswrapper[5002]: I1014 09:54:21.245784 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d8f34f80-d17b-4cdb-9834-f2e96de161ae" (UID: "d8f34f80-d17b-4cdb-9834-f2e96de161ae"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:54:21 crc kubenswrapper[5002]: I1014 09:54:21.257508 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d8f34f80-d17b-4cdb-9834-f2e96de161ae-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.126230 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 14 09:54:23 crc kubenswrapper[5002]: E1014 09:54:23.127284 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f34f80-d17b-4cdb-9834-f2e96de161ae" containerName="tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.127310 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f34f80-d17b-4cdb-9834-f2e96de161ae" containerName="tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.127688 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f34f80-d17b-4cdb-9834-f2e96de161ae" containerName="tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.128888 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.140390 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.201923 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dszbx\" (UniqueName: \"kubernetes.io/projected/96b65a05-43d2-4d66-b467-3c88b9ce57e4-kube-api-access-dszbx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.202001 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.303403 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dszbx\" (UniqueName: \"kubernetes.io/projected/96b65a05-43d2-4d66-b467-3c88b9ce57e4-kube-api-access-dszbx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.303475 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.304116 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.322863 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dszbx\" (UniqueName: \"kubernetes.io/projected/96b65a05-43d2-4d66-b467-3c88b9ce57e4-kube-api-access-dszbx\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.338215 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"96b65a05-43d2-4d66-b467-3c88b9ce57e4\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.470898 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Oct 14 09:54:23 crc kubenswrapper[5002]: I1014 09:54:23.982430 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Oct 14 09:54:24 crc kubenswrapper[5002]: I1014 09:54:24.092567 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"96b65a05-43d2-4d66-b467-3c88b9ce57e4","Type":"ContainerStarted","Data":"b90cb18ee09da5e631c0bcdcd656b96c54d77f5b4ce057a267526d1e7a70eba9"} Oct 14 09:54:25 crc kubenswrapper[5002]: I1014 09:54:25.109658 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"96b65a05-43d2-4d66-b467-3c88b9ce57e4","Type":"ContainerStarted","Data":"214e8120caca41315fbe4bef3789f7af1dd43f973e14e25841c36248084dba45"} Oct 14 09:54:25 crc kubenswrapper[5002]: I1014 09:54:25.143401 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=1.561886317 podStartE2EDuration="2.143380324s" podCreationTimestamp="2025-10-14 09:54:23 +0000 UTC" firstStartedPulling="2025-10-14 09:54:23.968828913 +0000 UTC m=+7396.950068395" lastFinishedPulling="2025-10-14 09:54:24.55032292 +0000 UTC m=+7397.531562402" observedRunningTime="2025-10-14 09:54:25.130749558 +0000 UTC m=+7398.111989020" watchObservedRunningTime="2025-10-14 09:54:25.143380324 +0000 UTC m=+7398.124619786" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.429755 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.431328 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.435685 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.436697 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.448693 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562248 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562308 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562391 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562424 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562459 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562496 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562519 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562616 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562680 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27m7t\" (UniqueName: \"kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.562857 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665178 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665268 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665305 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665410 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665513 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27m7t\" (UniqueName: \"kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665591 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665674 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.665716 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.666620 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.666697 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.667134 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.667485 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.667930 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.674002 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.674329 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.675701 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.676198 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.677982 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.694808 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.695342 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27m7t\" (UniqueName: \"kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.744512 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ansibletest-ansibletest\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " pod="openstack/ansibletest-ansibletest" Oct 14 09:54:42 crc kubenswrapper[5002]: I1014 09:54:42.756062 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 14 09:54:43 crc kubenswrapper[5002]: I1014 09:54:43.256383 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Oct 14 09:54:43 crc kubenswrapper[5002]: I1014 09:54:43.320225 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0d845551-0fcc-4cc0-8d67-7dedb66eabb5","Type":"ContainerStarted","Data":"4187f942fd637c451fc0e6b3b5e274deb729e2b5080c88be330d7b8b105fcd5b"} Oct 14 09:55:09 crc kubenswrapper[5002]: I1014 09:55:09.218786 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:55:09 crc kubenswrapper[5002]: I1014 09:55:09.219374 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:55:10 crc kubenswrapper[5002]: I1014 09:55:10.644667 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0d845551-0fcc-4cc0-8d67-7dedb66eabb5","Type":"ContainerStarted","Data":"4a4f744df87f2cf23ab234476913de80bfcbfcf9c2d98d5a4c897db9d7e6678f"} Oct 14 09:55:10 crc kubenswrapper[5002]: I1014 09:55:10.671627 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=4.005174048 podStartE2EDuration="29.671611252s" podCreationTimestamp="2025-10-14 09:54:41 +0000 UTC" firstStartedPulling="2025-10-14 09:54:43.261909382 +0000 UTC m=+7416.243148854" lastFinishedPulling="2025-10-14 09:55:08.928346566 +0000 UTC m=+7441.909586058" observedRunningTime="2025-10-14 09:55:10.671164381 +0000 UTC m=+7443.652403903" watchObservedRunningTime="2025-10-14 09:55:10.671611252 +0000 UTC m=+7443.652850704" Oct 14 09:55:11 crc kubenswrapper[5002]: I1014 09:55:11.660627 5002 generic.go:334] "Generic (PLEG): container finished" podID="0d845551-0fcc-4cc0-8d67-7dedb66eabb5" containerID="4a4f744df87f2cf23ab234476913de80bfcbfcf9c2d98d5a4c897db9d7e6678f" exitCode=0 Oct 14 09:55:11 crc kubenswrapper[5002]: I1014 09:55:11.660723 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0d845551-0fcc-4cc0-8d67-7dedb66eabb5","Type":"ContainerDied","Data":"4a4f744df87f2cf23ab234476913de80bfcbfcf9c2d98d5a4c897db9d7e6678f"} Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.036645 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163411 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163489 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163592 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27m7t\" (UniqueName: \"kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163647 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163705 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163780 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163818 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.163953 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.164111 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config\") pod \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\" (UID: \"0d845551-0fcc-4cc0-8d67-7dedb66eabb5\") " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.169591 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.170300 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph" (OuterVolumeSpecName: "ceph") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.172326 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t" (OuterVolumeSpecName: "kube-api-access-27m7t") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "kube-api-access-27m7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.180039 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.184432 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.196176 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.204394 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.215804 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.233253 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.236269 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0d845551-0fcc-4cc0-8d67-7dedb66eabb5" (UID: "0d845551-0fcc-4cc0-8d67-7dedb66eabb5"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267256 5002 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267298 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267325 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267339 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27m7t\" (UniqueName: \"kubernetes.io/projected/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-kube-api-access-27m7t\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267353 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267364 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267377 5002 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267390 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267421 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.267434 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d845551-0fcc-4cc0-8d67-7dedb66eabb5-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.293743 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.369538 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.691250 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"0d845551-0fcc-4cc0-8d67-7dedb66eabb5","Type":"ContainerDied","Data":"4187f942fd637c451fc0e6b3b5e274deb729e2b5080c88be330d7b8b105fcd5b"} Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.691300 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4187f942fd637c451fc0e6b3b5e274deb729e2b5080c88be330d7b8b105fcd5b" Oct 14 09:55:13 crc kubenswrapper[5002]: I1014 09:55:13.691574 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.504621 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 14 09:55:16 crc kubenswrapper[5002]: E1014 09:55:16.505716 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d845551-0fcc-4cc0-8d67-7dedb66eabb5" containerName="ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.505742 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d845551-0fcc-4cc0-8d67-7dedb66eabb5" containerName="ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.506150 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d845551-0fcc-4cc0-8d67-7dedb66eabb5" containerName="ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.507306 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.526728 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.645756 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.646461 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkvfr\" (UniqueName: \"kubernetes.io/projected/487e1aee-dbf2-451c-86dd-e2df3f229d89-kube-api-access-pkvfr\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.749332 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.749787 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkvfr\" (UniqueName: \"kubernetes.io/projected/487e1aee-dbf2-451c-86dd-e2df3f229d89-kube-api-access-pkvfr\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.750005 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.783395 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkvfr\" (UniqueName: \"kubernetes.io/projected/487e1aee-dbf2-451c-86dd-e2df3f229d89-kube-api-access-pkvfr\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.800504 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"487e1aee-dbf2-451c-86dd-e2df3f229d89\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:16 crc kubenswrapper[5002]: I1014 09:55:16.847748 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Oct 14 09:55:17 crc kubenswrapper[5002]: I1014 09:55:17.171207 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Oct 14 09:55:17 crc kubenswrapper[5002]: I1014 09:55:17.761768 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"487e1aee-dbf2-451c-86dd-e2df3f229d89","Type":"ContainerStarted","Data":"3dadfdfbb033f4075205248ee19381aa40c591adee3c1dd9b2a0231afdc2169c"} Oct 14 09:55:18 crc kubenswrapper[5002]: I1014 09:55:18.778177 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"487e1aee-dbf2-451c-86dd-e2df3f229d89","Type":"ContainerStarted","Data":"ea88959966bc88f6a5959d752d49d4ec5f87d52210847c6f075d8907b8c60f2f"} Oct 14 09:55:18 crc kubenswrapper[5002]: I1014 09:55:18.808056 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=2.366494995 podStartE2EDuration="2.808033561s" podCreationTimestamp="2025-10-14 09:55:16 +0000 UTC" firstStartedPulling="2025-10-14 09:55:17.167924887 +0000 UTC m=+7450.149164329" lastFinishedPulling="2025-10-14 09:55:17.609463443 +0000 UTC m=+7450.590702895" observedRunningTime="2025-10-14 09:55:18.802183316 +0000 UTC m=+7451.783422838" watchObservedRunningTime="2025-10-14 09:55:18.808033561 +0000 UTC m=+7451.789273043" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.016799 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.020080 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.023269 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.025744 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.062006 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.089787 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.090043 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192357 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192449 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192489 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192511 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192565 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192608 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192682 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8czg\" (UniqueName: \"kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.192892 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.193701 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.199136 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294243 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294337 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294358 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294399 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294457 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8czg\" (UniqueName: \"kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294514 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.294732 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.295231 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.295655 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.299784 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.304572 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.315549 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8czg\" (UniqueName: \"kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.334805 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"horizontest-tests-horizontest\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.345870 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 14 09:55:36 crc kubenswrapper[5002]: I1014 09:55:36.846701 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Oct 14 09:55:37 crc kubenswrapper[5002]: I1014 09:55:37.011583 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c","Type":"ContainerStarted","Data":"ea5b940f0061dc9ce9d6d09a4fae00ad054dfa0c56864279e865d97f92bcf4ea"} Oct 14 09:55:39 crc kubenswrapper[5002]: I1014 09:55:39.218024 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:55:39 crc kubenswrapper[5002]: I1014 09:55:39.218433 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.218629 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.219164 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.219215 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.219927 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.219991 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e" gracePeriod=600 Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.434543 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e" exitCode=0 Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.434604 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e"} Oct 14 09:56:09 crc kubenswrapper[5002]: I1014 09:56:09.434696 5002 scope.go:117] "RemoveContainer" containerID="19d656d90e146508f44e974b073d2133152627881dc71289428c19732fb159f5" Oct 14 09:56:12 crc kubenswrapper[5002]: E1014 09:56:12.222560 5002 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Oct 14 09:56:12 crc kubenswrapper[5002]: E1014 09:56:12.224884 5002 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r8czg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(df77ae37-bb4c-433c-bd13-bc48e1e6fa2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 09:56:12 crc kubenswrapper[5002]: E1014 09:56:12.226376 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" Oct 14 09:56:12 crc kubenswrapper[5002]: I1014 09:56:12.495500 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b"} Oct 14 09:56:12 crc kubenswrapper[5002]: E1014 09:56:12.497750 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.526203 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.530321 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.541959 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.639858 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9nk\" (UniqueName: \"kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.640282 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.640547 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.743420 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.743942 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.744247 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.744264 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9nk\" (UniqueName: \"kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.744649 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.772079 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9nk\" (UniqueName: \"kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk\") pod \"redhat-marketplace-jpl6t\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:14 crc kubenswrapper[5002]: I1014 09:56:14.876363 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:15 crc kubenswrapper[5002]: I1014 09:56:15.376156 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:15 crc kubenswrapper[5002]: I1014 09:56:15.528015 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerStarted","Data":"84ae33782c5e3177be1972e2a7f85c35f9fd548765949b32888529a862c3cc22"} Oct 14 09:56:16 crc kubenswrapper[5002]: I1014 09:56:16.548358 5002 generic.go:334] "Generic (PLEG): container finished" podID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerID="7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c" exitCode=0 Oct 14 09:56:16 crc kubenswrapper[5002]: I1014 09:56:16.548575 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerDied","Data":"7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c"} Oct 14 09:56:18 crc kubenswrapper[5002]: I1014 09:56:18.577016 5002 generic.go:334] "Generic (PLEG): container finished" podID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerID="bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844" exitCode=0 Oct 14 09:56:18 crc kubenswrapper[5002]: I1014 09:56:18.577186 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerDied","Data":"bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844"} Oct 14 09:56:18 crc kubenswrapper[5002]: I1014 09:56:18.581930 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:56:19 crc kubenswrapper[5002]: I1014 09:56:19.592682 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerStarted","Data":"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45"} Oct 14 09:56:19 crc kubenswrapper[5002]: I1014 09:56:19.622219 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jpl6t" podStartSLOduration=3.046707068 podStartE2EDuration="5.622196655s" podCreationTimestamp="2025-10-14 09:56:14 +0000 UTC" firstStartedPulling="2025-10-14 09:56:16.553985501 +0000 UTC m=+7509.535224963" lastFinishedPulling="2025-10-14 09:56:19.129475058 +0000 UTC m=+7512.110714550" observedRunningTime="2025-10-14 09:56:19.612872868 +0000 UTC m=+7512.594112350" watchObservedRunningTime="2025-10-14 09:56:19.622196655 +0000 UTC m=+7512.603436117" Oct 14 09:56:24 crc kubenswrapper[5002]: I1014 09:56:24.877217 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:24 crc kubenswrapper[5002]: I1014 09:56:24.877783 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:24 crc kubenswrapper[5002]: I1014 09:56:24.958247 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:25 crc kubenswrapper[5002]: I1014 09:56:25.655027 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c","Type":"ContainerStarted","Data":"e7aac4eb024b09c797dc0b2c401c6ddce10e40359097701c92633ab1eb2b5271"} Oct 14 09:56:25 crc kubenswrapper[5002]: I1014 09:56:25.699239 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=4.27801247 podStartE2EDuration="51.699214375s" podCreationTimestamp="2025-10-14 09:55:34 +0000 UTC" firstStartedPulling="2025-10-14 09:55:36.854357629 +0000 UTC m=+7469.835597121" lastFinishedPulling="2025-10-14 09:56:24.275559564 +0000 UTC m=+7517.256799026" observedRunningTime="2025-10-14 09:56:25.683701722 +0000 UTC m=+7518.664941184" watchObservedRunningTime="2025-10-14 09:56:25.699214375 +0000 UTC m=+7518.680453857" Oct 14 09:56:25 crc kubenswrapper[5002]: I1014 09:56:25.767903 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:25 crc kubenswrapper[5002]: I1014 09:56:25.838999 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:27 crc kubenswrapper[5002]: I1014 09:56:27.679285 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jpl6t" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="registry-server" containerID="cri-o://d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45" gracePeriod=2 Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.264862 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.431149 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content\") pod \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.431227 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf9nk\" (UniqueName: \"kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk\") pod \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.431397 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities\") pod \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\" (UID: \"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242\") " Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.432927 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities" (OuterVolumeSpecName: "utilities") pod "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" (UID: "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.441064 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk" (OuterVolumeSpecName: "kube-api-access-tf9nk") pod "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" (UID: "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242"). InnerVolumeSpecName "kube-api-access-tf9nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.452745 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" (UID: "e9c5cf2f-8f1c-4665-b6e8-37bf962e6242"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.535021 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.535101 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf9nk\" (UniqueName: \"kubernetes.io/projected/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-kube-api-access-tf9nk\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.535126 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.695336 5002 generic.go:334] "Generic (PLEG): container finished" podID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerID="d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45" exitCode=0 Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.695399 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerDied","Data":"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45"} Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.695418 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jpl6t" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.695437 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jpl6t" event={"ID":"e9c5cf2f-8f1c-4665-b6e8-37bf962e6242","Type":"ContainerDied","Data":"84ae33782c5e3177be1972e2a7f85c35f9fd548765949b32888529a862c3cc22"} Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.695468 5002 scope.go:117] "RemoveContainer" containerID="d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.746333 5002 scope.go:117] "RemoveContainer" containerID="bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.757737 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.772955 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jpl6t"] Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.794568 5002 scope.go:117] "RemoveContainer" containerID="7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.844177 5002 scope.go:117] "RemoveContainer" containerID="d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45" Oct 14 09:56:28 crc kubenswrapper[5002]: E1014 09:56:28.844797 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45\": container with ID starting with d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45 not found: ID does not exist" containerID="d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.844891 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45"} err="failed to get container status \"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45\": rpc error: code = NotFound desc = could not find container \"d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45\": container with ID starting with d0f4d1152f731ac73cc9509f34a9d0af5a7c159f95741cdbe86a2cbf047f9d45 not found: ID does not exist" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.844933 5002 scope.go:117] "RemoveContainer" containerID="bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844" Oct 14 09:56:28 crc kubenswrapper[5002]: E1014 09:56:28.845476 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844\": container with ID starting with bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844 not found: ID does not exist" containerID="bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.845526 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844"} err="failed to get container status \"bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844\": rpc error: code = NotFound desc = could not find container \"bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844\": container with ID starting with bfb00d5807a9bf54d9ab3a11ffb34c10bd072a10055edb0d05a8047b8951d844 not found: ID does not exist" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.845560 5002 scope.go:117] "RemoveContainer" containerID="7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c" Oct 14 09:56:28 crc kubenswrapper[5002]: E1014 09:56:28.845936 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c\": container with ID starting with 7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c not found: ID does not exist" containerID="7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c" Oct 14 09:56:28 crc kubenswrapper[5002]: I1014 09:56:28.845986 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c"} err="failed to get container status \"7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c\": rpc error: code = NotFound desc = could not find container \"7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c\": container with ID starting with 7df32df35143c8d6283b6c91375770871b97ab40bfc7ea8066789a4e4572209c not found: ID does not exist" Oct 14 09:56:29 crc kubenswrapper[5002]: I1014 09:56:29.740991 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" path="/var/lib/kubelet/pods/e9c5cf2f-8f1c-4665-b6e8-37bf962e6242/volumes" Oct 14 09:58:24 crc kubenswrapper[5002]: I1014 09:58:24.108486 5002 generic.go:334] "Generic (PLEG): container finished" podID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" containerID="e7aac4eb024b09c797dc0b2c401c6ddce10e40359097701c92633ab1eb2b5271" exitCode=0 Oct 14 09:58:24 crc kubenswrapper[5002]: I1014 09:58:24.108643 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c","Type":"ContainerDied","Data":"e7aac4eb024b09c797dc0b2c401c6ddce10e40359097701c92633ab1eb2b5271"} Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.555796 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:25 crc kubenswrapper[5002]: E1014 09:58:25.556527 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="extract-utilities" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.556543 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="extract-utilities" Oct 14 09:58:25 crc kubenswrapper[5002]: E1014 09:58:25.556567 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="extract-content" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.556576 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="extract-content" Oct 14 09:58:25 crc kubenswrapper[5002]: E1014 09:58:25.556601 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="registry-server" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.556609 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="registry-server" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.556926 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c5cf2f-8f1c-4665-b6e8-37bf962e6242" containerName="registry-server" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.558771 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.566060 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.608583 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.689914 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690085 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690142 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690164 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690224 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690254 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690326 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8czg\" (UniqueName: \"kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690383 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config\") pod \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\" (UID: \"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c\") " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690644 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690764 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smvp8\" (UniqueName: \"kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690787 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.690979 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.691317 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.698512 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph" (OuterVolumeSpecName: "ceph") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.698525 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.698990 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg" (OuterVolumeSpecName: "kube-api-access-r8czg") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "kube-api-access-r8czg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.740047 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.786639 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.815748 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.816036 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.821724 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smvp8\" (UniqueName: \"kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.821923 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.821955 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.823436 5002 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.823467 5002 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.823481 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8czg\" (UniqueName: \"kubernetes.io/projected/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-kube-api-access-r8czg\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.823495 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.824315 5002 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.831165 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:25 crc kubenswrapper[5002]: E1014 09:58:25.831601 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" containerName="horizontest-tests-horizontest" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.831621 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" containerName="horizontest-tests-horizontest" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.831987 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" containerName="horizontest-tests-horizontest" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.836280 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.836440 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.844072 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smvp8\" (UniqueName: \"kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8\") pod \"certified-operators-rqcsw\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.845076 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.857085 5002 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.923109 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.925645 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.925886 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.926121 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl8hd\" (UniqueName: \"kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.926223 5002 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:25 crc kubenswrapper[5002]: I1014 09:58:25.926279 5002 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.002204 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c" (UID: "df77ae37-bb4c-433c-bd13-bc48e1e6fa2c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.027386 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl8hd\" (UniqueName: \"kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.027433 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.027519 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.027572 5002 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/df77ae37-bb4c-433c-bd13-bc48e1e6fa2c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.027972 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.028426 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.050825 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl8hd\" (UniqueName: \"kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd\") pod \"redhat-operators-qn4ql\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.155754 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"df77ae37-bb4c-433c-bd13-bc48e1e6fa2c","Type":"ContainerDied","Data":"ea5b940f0061dc9ce9d6d09a4fae00ad054dfa0c56864279e865d97f92bcf4ea"} Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.155792 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea5b940f0061dc9ce9d6d09a4fae00ad054dfa0c56864279e865d97f92bcf4ea" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.155890 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.173361 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.224852 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:26 crc kubenswrapper[5002]: I1014 09:58:26.633525 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.166001 5002 generic.go:334] "Generic (PLEG): container finished" podID="4272268f-203d-4145-870e-14b38460a871" containerID="3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404" exitCode=0 Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.166074 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerDied","Data":"3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404"} Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.166378 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerStarted","Data":"010b2ab27e4d8366946963232b0ea4a86ae72d209d4b942501c0642db635afb0"} Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.170069 5002 generic.go:334] "Generic (PLEG): container finished" podID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerID="0bd7fdab91b95635bdce741696e4e9e28bd274f2884683c65d6d5c2a43914129" exitCode=0 Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.170117 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerDied","Data":"0bd7fdab91b95635bdce741696e4e9e28bd274f2884683c65d6d5c2a43914129"} Oct 14 09:58:27 crc kubenswrapper[5002]: I1014 09:58:27.170151 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerStarted","Data":"ffbdb3c123de64df2305c8752169b72b2ddf92fab2b9525011a5fdca2745f5f1"} Oct 14 09:58:28 crc kubenswrapper[5002]: I1014 09:58:28.182248 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerStarted","Data":"75e456dcb310f3ad73f425d81ceea516cc765ea15e82f4faed04072da220d778"} Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.195104 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerStarted","Data":"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8"} Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.702954 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.707407 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.747263 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.814678 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.815307 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psq4j\" (UniqueName: \"kubernetes.io/projected/845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1-kube-api-access-psq4j\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.917011 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psq4j\" (UniqueName: \"kubernetes.io/projected/845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1-kube-api-access-psq4j\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.917101 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.918081 5002 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.944203 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psq4j\" (UniqueName: \"kubernetes.io/projected/845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1-kube-api-access-psq4j\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:29 crc kubenswrapper[5002]: I1014 09:58:29.955042 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:30 crc kubenswrapper[5002]: I1014 09:58:30.039573 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Oct 14 09:58:30 crc kubenswrapper[5002]: E1014 09:58:30.039814 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 09:58:30 crc kubenswrapper[5002]: I1014 09:58:30.504545 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Oct 14 09:58:30 crc kubenswrapper[5002]: W1014 09:58:30.529206 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845f0c8f_dc7f_4de3_8b0d_6fd3428c15f1.slice/crio-8db8409105fa8eb2d387e843cceb771b699eaa26497f3643e2d9c72098a4607a WatchSource:0}: Error finding container 8db8409105fa8eb2d387e843cceb771b699eaa26497f3643e2d9c72098a4607a: Status 404 returned error can't find the container with id 8db8409105fa8eb2d387e843cceb771b699eaa26497f3643e2d9c72098a4607a Oct 14 09:58:30 crc kubenswrapper[5002]: E1014 09:58:30.529938 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 09:58:31 crc kubenswrapper[5002]: E1014 09:58:31.124273 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 09:58:31 crc kubenswrapper[5002]: I1014 09:58:31.227201 5002 generic.go:334] "Generic (PLEG): container finished" podID="4272268f-203d-4145-870e-14b38460a871" containerID="ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8" exitCode=0 Oct 14 09:58:31 crc kubenswrapper[5002]: I1014 09:58:31.227310 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerDied","Data":"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8"} Oct 14 09:58:31 crc kubenswrapper[5002]: I1014 09:58:31.232315 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1","Type":"ContainerStarted","Data":"8db8409105fa8eb2d387e843cceb771b699eaa26497f3643e2d9c72098a4607a"} Oct 14 09:58:31 crc kubenswrapper[5002]: I1014 09:58:31.237386 5002 generic.go:334] "Generic (PLEG): container finished" podID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerID="75e456dcb310f3ad73f425d81ceea516cc765ea15e82f4faed04072da220d778" exitCode=0 Oct 14 09:58:31 crc kubenswrapper[5002]: I1014 09:58:31.237441 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerDied","Data":"75e456dcb310f3ad73f425d81ceea516cc765ea15e82f4faed04072da220d778"} Oct 14 09:58:32 crc kubenswrapper[5002]: I1014 09:58:32.275073 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerStarted","Data":"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86"} Oct 14 09:58:32 crc kubenswrapper[5002]: I1014 09:58:32.285279 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1","Type":"ContainerStarted","Data":"332f2d859ca64f4c608441025b19839018465b92b3f72f396fed0ecb6dbc6c11"} Oct 14 09:58:32 crc kubenswrapper[5002]: E1014 09:58:32.285997 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 09:58:32 crc kubenswrapper[5002]: I1014 09:58:32.307240 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rqcsw" podStartSLOduration=2.604083386 podStartE2EDuration="7.307217177s" podCreationTimestamp="2025-10-14 09:58:25 +0000 UTC" firstStartedPulling="2025-10-14 09:58:27.167621564 +0000 UTC m=+7640.148861006" lastFinishedPulling="2025-10-14 09:58:31.870755305 +0000 UTC m=+7644.851994797" observedRunningTime="2025-10-14 09:58:32.301529904 +0000 UTC m=+7645.282769376" watchObservedRunningTime="2025-10-14 09:58:32.307217177 +0000 UTC m=+7645.288456639" Oct 14 09:58:32 crc kubenswrapper[5002]: I1014 09:58:32.328620 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=2.735351456 podStartE2EDuration="3.328571524s" podCreationTimestamp="2025-10-14 09:58:29 +0000 UTC" firstStartedPulling="2025-10-14 09:58:30.530971243 +0000 UTC m=+7643.512210705" lastFinishedPulling="2025-10-14 09:58:31.124191281 +0000 UTC m=+7644.105430773" observedRunningTime="2025-10-14 09:58:32.317188491 +0000 UTC m=+7645.298427983" watchObservedRunningTime="2025-10-14 09:58:32.328571524 +0000 UTC m=+7645.309810986" Oct 14 09:58:33 crc kubenswrapper[5002]: I1014 09:58:33.310169 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerStarted","Data":"95922fdbe382c96e381df08f17f8e27eed8d2f2e897625d52b96ecdd601e3843"} Oct 14 09:58:33 crc kubenswrapper[5002]: E1014 09:58:33.311038 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 09:58:33 crc kubenswrapper[5002]: I1014 09:58:33.366442 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qn4ql" podStartSLOduration=3.559424407 podStartE2EDuration="8.36642483s" podCreationTimestamp="2025-10-14 09:58:25 +0000 UTC" firstStartedPulling="2025-10-14 09:58:27.17127192 +0000 UTC m=+7640.152511372" lastFinishedPulling="2025-10-14 09:58:31.978272343 +0000 UTC m=+7644.959511795" observedRunningTime="2025-10-14 09:58:33.3618935 +0000 UTC m=+7646.343133012" watchObservedRunningTime="2025-10-14 09:58:33.36642483 +0000 UTC m=+7646.347664282" Oct 14 09:58:35 crc kubenswrapper[5002]: I1014 09:58:35.924294 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:35 crc kubenswrapper[5002]: I1014 09:58:35.924703 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:35 crc kubenswrapper[5002]: I1014 09:58:35.970993 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:36 crc kubenswrapper[5002]: I1014 09:58:36.174250 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:36 crc kubenswrapper[5002]: I1014 09:58:36.174288 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:37 crc kubenswrapper[5002]: I1014 09:58:37.240977 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qn4ql" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="registry-server" probeResult="failure" output=< Oct 14 09:58:37 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 09:58:37 crc kubenswrapper[5002]: > Oct 14 09:58:39 crc kubenswrapper[5002]: I1014 09:58:39.218111 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:58:39 crc kubenswrapper[5002]: I1014 09:58:39.218510 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.007412 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.147080 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.213609 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.268234 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.450619 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rqcsw" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="registry-server" containerID="cri-o://ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86" gracePeriod=2 Oct 14 09:58:46 crc kubenswrapper[5002]: I1014 09:58:46.999519 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.149688 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smvp8\" (UniqueName: \"kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8\") pod \"4272268f-203d-4145-870e-14b38460a871\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.150006 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities\") pod \"4272268f-203d-4145-870e-14b38460a871\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.150108 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content\") pod \"4272268f-203d-4145-870e-14b38460a871\" (UID: \"4272268f-203d-4145-870e-14b38460a871\") " Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.151113 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities" (OuterVolumeSpecName: "utilities") pod "4272268f-203d-4145-870e-14b38460a871" (UID: "4272268f-203d-4145-870e-14b38460a871"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.156486 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8" (OuterVolumeSpecName: "kube-api-access-smvp8") pod "4272268f-203d-4145-870e-14b38460a871" (UID: "4272268f-203d-4145-870e-14b38460a871"). InnerVolumeSpecName "kube-api-access-smvp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.204085 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4272268f-203d-4145-870e-14b38460a871" (UID: "4272268f-203d-4145-870e-14b38460a871"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.252925 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.252957 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4272268f-203d-4145-870e-14b38460a871-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.252975 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smvp8\" (UniqueName: \"kubernetes.io/projected/4272268f-203d-4145-870e-14b38460a871-kube-api-access-smvp8\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.468200 5002 generic.go:334] "Generic (PLEG): container finished" podID="4272268f-203d-4145-870e-14b38460a871" containerID="ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86" exitCode=0 Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.468286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerDied","Data":"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86"} Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.468359 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rqcsw" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.468678 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rqcsw" event={"ID":"4272268f-203d-4145-870e-14b38460a871","Type":"ContainerDied","Data":"010b2ab27e4d8366946963232b0ea4a86ae72d209d4b942501c0642db635afb0"} Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.468921 5002 scope.go:117] "RemoveContainer" containerID="ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.514095 5002 scope.go:117] "RemoveContainer" containerID="ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.518417 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.527906 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rqcsw"] Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.545525 5002 scope.go:117] "RemoveContainer" containerID="3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.617553 5002 scope.go:117] "RemoveContainer" containerID="ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86" Oct 14 09:58:47 crc kubenswrapper[5002]: E1014 09:58:47.618021 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86\": container with ID starting with ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86 not found: ID does not exist" containerID="ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.618051 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86"} err="failed to get container status \"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86\": rpc error: code = NotFound desc = could not find container \"ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86\": container with ID starting with ddb267cd0fd9f371fa33b76011e755628891d6236d0e55745714374ad78d6e86 not found: ID does not exist" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.618071 5002 scope.go:117] "RemoveContainer" containerID="ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8" Oct 14 09:58:47 crc kubenswrapper[5002]: E1014 09:58:47.618626 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8\": container with ID starting with ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8 not found: ID does not exist" containerID="ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.618650 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8"} err="failed to get container status \"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8\": rpc error: code = NotFound desc = could not find container \"ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8\": container with ID starting with ededd6d0ca05820ddae14706cc1cf0e9680edb9a8919772ebedda1b3c11254a8 not found: ID does not exist" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.618670 5002 scope.go:117] "RemoveContainer" containerID="3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404" Oct 14 09:58:47 crc kubenswrapper[5002]: E1014 09:58:47.618934 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404\": container with ID starting with 3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404 not found: ID does not exist" containerID="3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.618955 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404"} err="failed to get container status \"3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404\": rpc error: code = NotFound desc = could not find container \"3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404\": container with ID starting with 3f204dc16b1e66bbdbaa7bc79a687cfd3f96e33e70b489b5c9c21ef5a1df7404 not found: ID does not exist" Oct 14 09:58:47 crc kubenswrapper[5002]: I1014 09:58:47.734059 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4272268f-203d-4145-870e-14b38460a871" path="/var/lib/kubelet/pods/4272268f-203d-4145-870e-14b38460a871/volumes" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.263706 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.264414 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qn4ql" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="registry-server" containerID="cri-o://95922fdbe382c96e381df08f17f8e27eed8d2f2e897625d52b96ecdd601e3843" gracePeriod=2 Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.483779 5002 generic.go:334] "Generic (PLEG): container finished" podID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerID="95922fdbe382c96e381df08f17f8e27eed8d2f2e897625d52b96ecdd601e3843" exitCode=0 Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.483877 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerDied","Data":"95922fdbe382c96e381df08f17f8e27eed8d2f2e897625d52b96ecdd601e3843"} Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.777774 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.885801 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content\") pod \"5512263b-59a2-4b6c-b7f7-13f374363c7b\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.886001 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl8hd\" (UniqueName: \"kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd\") pod \"5512263b-59a2-4b6c-b7f7-13f374363c7b\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.886061 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities\") pod \"5512263b-59a2-4b6c-b7f7-13f374363c7b\" (UID: \"5512263b-59a2-4b6c-b7f7-13f374363c7b\") " Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.886901 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities" (OuterVolumeSpecName: "utilities") pod "5512263b-59a2-4b6c-b7f7-13f374363c7b" (UID: "5512263b-59a2-4b6c-b7f7-13f374363c7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.898172 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd" (OuterVolumeSpecName: "kube-api-access-pl8hd") pod "5512263b-59a2-4b6c-b7f7-13f374363c7b" (UID: "5512263b-59a2-4b6c-b7f7-13f374363c7b"). InnerVolumeSpecName "kube-api-access-pl8hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.978088 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5512263b-59a2-4b6c-b7f7-13f374363c7b" (UID: "5512263b-59a2-4b6c-b7f7-13f374363c7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.988342 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.988385 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl8hd\" (UniqueName: \"kubernetes.io/projected/5512263b-59a2-4b6c-b7f7-13f374363c7b-kube-api-access-pl8hd\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:48 crc kubenswrapper[5002]: I1014 09:58:48.988405 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5512263b-59a2-4b6c-b7f7-13f374363c7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.504380 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qn4ql" event={"ID":"5512263b-59a2-4b6c-b7f7-13f374363c7b","Type":"ContainerDied","Data":"ffbdb3c123de64df2305c8752169b72b2ddf92fab2b9525011a5fdca2745f5f1"} Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.504459 5002 scope.go:117] "RemoveContainer" containerID="95922fdbe382c96e381df08f17f8e27eed8d2f2e897625d52b96ecdd601e3843" Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.504541 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qn4ql" Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.562665 5002 scope.go:117] "RemoveContainer" containerID="75e456dcb310f3ad73f425d81ceea516cc765ea15e82f4faed04072da220d778" Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.573468 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.585147 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qn4ql"] Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.598259 5002 scope.go:117] "RemoveContainer" containerID="0bd7fdab91b95635bdce741696e4e9e28bd274f2884683c65d6d5c2a43914129" Oct 14 09:58:49 crc kubenswrapper[5002]: I1014 09:58:49.739757 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" path="/var/lib/kubelet/pods/5512263b-59a2-4b6c-b7f7-13f374363c7b/volumes" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.292806 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pl9xf/must-gather-jsjrs"] Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293638 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="extract-utilities" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293657 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="extract-utilities" Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293669 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293676 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293695 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="extract-content" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293703 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="extract-content" Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293729 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293739 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293763 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="extract-content" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293770 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="extract-content" Oct 14 09:58:51 crc kubenswrapper[5002]: E1014 09:58:51.293794 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="extract-utilities" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.293802 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="extract-utilities" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.294055 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="5512263b-59a2-4b6c-b7f7-13f374363c7b" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.294074 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4272268f-203d-4145-870e-14b38460a871" containerName="registry-server" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.295323 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.297430 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pl9xf"/"openshift-service-ca.crt" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.297679 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pl9xf"/"kube-root-ca.crt" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.297817 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pl9xf"/"default-dockercfg-48m2z" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.316029 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pl9xf/must-gather-jsjrs"] Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.434020 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gxs4\" (UniqueName: \"kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.434097 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.535689 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gxs4\" (UniqueName: \"kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.535784 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.536257 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.557952 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gxs4\" (UniqueName: \"kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4\") pod \"must-gather-jsjrs\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:51 crc kubenswrapper[5002]: I1014 09:58:51.626822 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 09:58:52 crc kubenswrapper[5002]: I1014 09:58:52.139658 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pl9xf/must-gather-jsjrs"] Oct 14 09:58:52 crc kubenswrapper[5002]: I1014 09:58:52.547378 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" event={"ID":"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46","Type":"ContainerStarted","Data":"3e35d30d8273afca9e75db4eefca5ebccdfc0a8b5fdf4cb0c481af41890d08f5"} Oct 14 09:58:58 crc kubenswrapper[5002]: I1014 09:58:58.601286 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" event={"ID":"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46","Type":"ContainerStarted","Data":"efa175bf4460d11990a9d5970622695b7b24c7e79cea897d117f0503b90e5d9b"} Oct 14 09:58:58 crc kubenswrapper[5002]: I1014 09:58:58.602027 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" event={"ID":"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46","Type":"ContainerStarted","Data":"b2e2073c1d98ef99ac374b68b0ae32b00728f6b231aceea1ac9d06dd3cbc5f9e"} Oct 14 09:58:58 crc kubenswrapper[5002]: I1014 09:58:58.620126 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" podStartSLOduration=2.1627467129999998 podStartE2EDuration="7.620099232s" podCreationTimestamp="2025-10-14 09:58:51 +0000 UTC" firstStartedPulling="2025-10-14 09:58:52.159367654 +0000 UTC m=+7665.140607136" lastFinishedPulling="2025-10-14 09:58:57.616720203 +0000 UTC m=+7670.597959655" observedRunningTime="2025-10-14 09:58:58.614874184 +0000 UTC m=+7671.596113656" watchObservedRunningTime="2025-10-14 09:58:58.620099232 +0000 UTC m=+7671.601338724" Oct 14 09:59:02 crc kubenswrapper[5002]: E1014 09:59:02.421110 5002 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.74:47536->38.102.83.74:44515: write tcp 38.102.83.74:47536->38.102.83.74:44515: write: broken pipe Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.596631 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-t6lfl"] Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.598101 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.721849 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.722259 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rddc5\" (UniqueName: \"kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.824403 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.824501 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rddc5\" (UniqueName: \"kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.824505 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.855423 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rddc5\" (UniqueName: \"kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5\") pod \"crc-debug-t6lfl\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:03 crc kubenswrapper[5002]: I1014 09:59:03.916112 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 09:59:04 crc kubenswrapper[5002]: I1014 09:59:04.665734 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" event={"ID":"c97817dc-4158-4066-b751-f887c2990852","Type":"ContainerStarted","Data":"130394c8279ac044e03d16b60e4caa781e8c26934a93fc53130b3427436f01e2"} Oct 14 09:59:09 crc kubenswrapper[5002]: I1014 09:59:09.218317 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:59:09 crc kubenswrapper[5002]: I1014 09:59:09.218873 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:59:14 crc kubenswrapper[5002]: I1014 09:59:14.758892 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" event={"ID":"c97817dc-4158-4066-b751-f887c2990852","Type":"ContainerStarted","Data":"87d2b0c7f8beb4f04ec984cd6147ee5c5e551abb0543e5165b638bcbc0c810d4"} Oct 14 09:59:14 crc kubenswrapper[5002]: I1014 09:59:14.782424 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" podStartSLOduration=1.465055535 podStartE2EDuration="11.782399543s" podCreationTimestamp="2025-10-14 09:59:03 +0000 UTC" firstStartedPulling="2025-10-14 09:59:03.949319366 +0000 UTC m=+7676.930558838" lastFinishedPulling="2025-10-14 09:59:14.266663394 +0000 UTC m=+7687.247902846" observedRunningTime="2025-10-14 09:59:14.771154334 +0000 UTC m=+7687.752393786" watchObservedRunningTime="2025-10-14 09:59:14.782399543 +0000 UTC m=+7687.763638995" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.218559 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.219147 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.219198 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.220014 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.220069 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" gracePeriod=600 Oct 14 09:59:39 crc kubenswrapper[5002]: E1014 09:59:39.345263 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.993788 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" exitCode=0 Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.993829 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b"} Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.993928 5002 scope.go:117] "RemoveContainer" containerID="07e1114a85ef339ea0bd87605ed059e48ab4408b1b2e7cdcf79c7827ed4d1e9e" Oct 14 09:59:39 crc kubenswrapper[5002]: I1014 09:59:39.995677 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 09:59:39 crc kubenswrapper[5002]: E1014 09:59:39.996315 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 09:59:50 crc kubenswrapper[5002]: I1014 09:59:50.720308 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 09:59:50 crc kubenswrapper[5002]: E1014 09:59:50.721182 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.169124 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh"] Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.173665 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.176289 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.177628 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.191333 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.191415 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtctw\" (UniqueName: \"kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.191506 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.192220 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh"] Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.293067 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtctw\" (UniqueName: \"kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.293382 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.293439 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.294706 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.307290 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.326936 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtctw\" (UniqueName: \"kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw\") pod \"collect-profiles-29340600-zwvbh\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.497227 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:00 crc kubenswrapper[5002]: I1014 10:00:00.825086 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh"] Oct 14 10:00:01 crc kubenswrapper[5002]: I1014 10:00:01.179959 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" event={"ID":"3d8b5e3b-f2af-4c47-93e3-13d1657224a3","Type":"ContainerStarted","Data":"546faa7b4a5fa719dcdca267115467b143083049af4e0503109e1259b910a4fa"} Oct 14 10:00:01 crc kubenswrapper[5002]: I1014 10:00:01.180276 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" event={"ID":"3d8b5e3b-f2af-4c47-93e3-13d1657224a3","Type":"ContainerStarted","Data":"5c2ae6201169dc3791dbb47a2ff74ccc6073c4938316c1c96abeb2657a48e7d9"} Oct 14 10:00:01 crc kubenswrapper[5002]: I1014 10:00:01.202649 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" podStartSLOduration=1.202622222 podStartE2EDuration="1.202622222s" podCreationTimestamp="2025-10-14 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:00:01.196876019 +0000 UTC m=+7734.178115491" watchObservedRunningTime="2025-10-14 10:00:01.202622222 +0000 UTC m=+7734.183861694" Oct 14 10:00:01 crc kubenswrapper[5002]: E1014 10:00:01.721330 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:00:02 crc kubenswrapper[5002]: I1014 10:00:02.197529 5002 generic.go:334] "Generic (PLEG): container finished" podID="3d8b5e3b-f2af-4c47-93e3-13d1657224a3" containerID="546faa7b4a5fa719dcdca267115467b143083049af4e0503109e1259b910a4fa" exitCode=0 Oct 14 10:00:02 crc kubenswrapper[5002]: I1014 10:00:02.197786 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" event={"ID":"3d8b5e3b-f2af-4c47-93e3-13d1657224a3","Type":"ContainerDied","Data":"546faa7b4a5fa719dcdca267115467b143083049af4e0503109e1259b910a4fa"} Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.580074 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.689299 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume\") pod \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.689428 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtctw\" (UniqueName: \"kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw\") pod \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.689613 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume\") pod \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\" (UID: \"3d8b5e3b-f2af-4c47-93e3-13d1657224a3\") " Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.690269 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d8b5e3b-f2af-4c47-93e3-13d1657224a3" (UID: "3d8b5e3b-f2af-4c47-93e3-13d1657224a3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.696624 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d8b5e3b-f2af-4c47-93e3-13d1657224a3" (UID: "3d8b5e3b-f2af-4c47-93e3-13d1657224a3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.704428 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw" (OuterVolumeSpecName: "kube-api-access-rtctw") pod "3d8b5e3b-f2af-4c47-93e3-13d1657224a3" (UID: "3d8b5e3b-f2af-4c47-93e3-13d1657224a3"). InnerVolumeSpecName "kube-api-access-rtctw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.792064 5002 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.792103 5002 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:03 crc kubenswrapper[5002]: I1014 10:00:03.792116 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtctw\" (UniqueName: \"kubernetes.io/projected/3d8b5e3b-f2af-4c47-93e3-13d1657224a3-kube-api-access-rtctw\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.214401 5002 generic.go:334] "Generic (PLEG): container finished" podID="c97817dc-4158-4066-b751-f887c2990852" containerID="87d2b0c7f8beb4f04ec984cd6147ee5c5e551abb0543e5165b638bcbc0c810d4" exitCode=0 Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.214467 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" event={"ID":"c97817dc-4158-4066-b751-f887c2990852","Type":"ContainerDied","Data":"87d2b0c7f8beb4f04ec984cd6147ee5c5e551abb0543e5165b638bcbc0c810d4"} Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.215944 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" event={"ID":"3d8b5e3b-f2af-4c47-93e3-13d1657224a3","Type":"ContainerDied","Data":"5c2ae6201169dc3791dbb47a2ff74ccc6073c4938316c1c96abeb2657a48e7d9"} Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.215976 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c2ae6201169dc3791dbb47a2ff74ccc6073c4938316c1c96abeb2657a48e7d9" Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.215982 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-zwvbh" Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.269195 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg"] Oct 14 10:00:04 crc kubenswrapper[5002]: I1014 10:00:04.276709 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-hpxzg"] Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.331666 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.370669 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-t6lfl"] Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.379869 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-t6lfl"] Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.529435 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host\") pod \"c97817dc-4158-4066-b751-f887c2990852\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.529554 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host" (OuterVolumeSpecName: "host") pod "c97817dc-4158-4066-b751-f887c2990852" (UID: "c97817dc-4158-4066-b751-f887c2990852"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.530139 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rddc5\" (UniqueName: \"kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5\") pod \"c97817dc-4158-4066-b751-f887c2990852\" (UID: \"c97817dc-4158-4066-b751-f887c2990852\") " Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.530806 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c97817dc-4158-4066-b751-f887c2990852-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.536345 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5" (OuterVolumeSpecName: "kube-api-access-rddc5") pod "c97817dc-4158-4066-b751-f887c2990852" (UID: "c97817dc-4158-4066-b751-f887c2990852"). InnerVolumeSpecName "kube-api-access-rddc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.632495 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rddc5\" (UniqueName: \"kubernetes.io/projected/c97817dc-4158-4066-b751-f887c2990852-kube-api-access-rddc5\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.721312 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:00:05 crc kubenswrapper[5002]: E1014 10:00:05.721920 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.750987 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c97817dc-4158-4066-b751-f887c2990852" path="/var/lib/kubelet/pods/c97817dc-4158-4066-b751-f887c2990852/volumes" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.752691 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc7a8845-3a0c-4f98-9605-6547d7a52746" path="/var/lib/kubelet/pods/fc7a8845-3a0c-4f98-9605-6547d7a52746/volumes" Oct 14 10:00:05 crc kubenswrapper[5002]: I1014 10:00:05.896791 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_0d845551-0fcc-4cc0-8d67-7dedb66eabb5/ansibletest-ansibletest/0.log" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.134282 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67d6789ffb-z5pvg_e3b04e6e-b1f7-4327-985d-6ed4e526eeaa/barbican-api/0.log" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.241554 5002 scope.go:117] "RemoveContainer" containerID="87d2b0c7f8beb4f04ec984cd6147ee5c5e551abb0543e5165b638bcbc0c810d4" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.241640 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-t6lfl" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.328873 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67d6789ffb-z5pvg_e3b04e6e-b1f7-4327-985d-6ed4e526eeaa/barbican-api-log/0.log" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.499133 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f846844d6-pvmrs_a1f84982-1a7e-4054-b3f5-666de26a4e96/barbican-keystone-listener/0.log" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.532623 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-nnd2v"] Oct 14 10:00:06 crc kubenswrapper[5002]: E1014 10:00:06.533143 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d8b5e3b-f2af-4c47-93e3-13d1657224a3" containerName="collect-profiles" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.533168 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d8b5e3b-f2af-4c47-93e3-13d1657224a3" containerName="collect-profiles" Oct 14 10:00:06 crc kubenswrapper[5002]: E1014 10:00:06.533194 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c97817dc-4158-4066-b751-f887c2990852" containerName="container-00" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.533204 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="c97817dc-4158-4066-b751-f887c2990852" containerName="container-00" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.533447 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="c97817dc-4158-4066-b751-f887c2990852" containerName="container-00" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.533485 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d8b5e3b-f2af-4c47-93e3-13d1657224a3" containerName="collect-profiles" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.534246 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.655443 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.655797 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m4gd\" (UniqueName: \"kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.757871 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m4gd\" (UniqueName: \"kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.758356 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.758483 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.775614 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m4gd\" (UniqueName: \"kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd\") pod \"crc-debug-nnd2v\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:06 crc kubenswrapper[5002]: I1014 10:00:06.850697 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.019982 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7fffd76679-n928f_ec0b4c68-f07d-4d42-8225-6950e9315970/barbican-worker/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.186076 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f846844d6-pvmrs_a1f84982-1a7e-4054-b3f5-666de26a4e96/barbican-keystone-listener-log/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.201105 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7fffd76679-n928f_ec0b4c68-f07d-4d42-8225-6950e9315970/barbican-worker-log/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.258465 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" event={"ID":"6f8824b5-d767-4e10-bf8c-32baee90c44b","Type":"ContainerStarted","Data":"fd2ba8e7a6c433d9bb274d54a086342774d51273612a3a09527953a6b620c988"} Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.259044 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" event={"ID":"6f8824b5-d767-4e10-bf8c-32baee90c44b","Type":"ContainerStarted","Data":"cfafc96db089f239cb8eb1489d3c2b7cf9e336730082b3ac9bd73c9f7387aec3"} Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.275288 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" podStartSLOduration=1.275268715 podStartE2EDuration="1.275268715s" podCreationTimestamp="2025-10-14 10:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:00:07.271885494 +0000 UTC m=+7740.253124966" watchObservedRunningTime="2025-10-14 10:00:07.275268715 +0000 UTC m=+7740.256508167" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.485355 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk_938c7955-b34d-435c-8933-2c3dc8043578/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.762126 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/ceilometer-central-agent/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.932366 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/ceilometer-notification-agent/0.log" Oct 14 10:00:07 crc kubenswrapper[5002]: I1014 10:00:07.956689 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/proxy-httpd/0.log" Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.141366 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/sg-core/0.log" Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.277115 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f8824b5-d767-4e10-bf8c-32baee90c44b" containerID="fd2ba8e7a6c433d9bb274d54a086342774d51273612a3a09527953a6b620c988" exitCode=0 Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.277151 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" event={"ID":"6f8824b5-d767-4e10-bf8c-32baee90c44b","Type":"ContainerDied","Data":"fd2ba8e7a6c433d9bb274d54a086342774d51273612a3a09527953a6b620c988"} Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.377776 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r_d93bac7d-9b37-4cfe-98dd-271aeff235ff/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.635412 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h_b871be8d-75a8-4c47-bfe4-4ad8f79d7fec/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:08 crc kubenswrapper[5002]: I1014 10:00:08.855924 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d29ae9ac-13f1-4621-9696-4eabea39109f/cinder-api-log/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.051351 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d29ae9ac-13f1-4621-9696-4eabea39109f/cinder-api/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.408242 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.412143 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_947933b0-5baf-4203-bea2-3b229c31ab91/cinder-backup/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.425647 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_947933b0-5baf-4203-bea2-3b229c31ab91/probe/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.511767 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host\") pod \"6f8824b5-d767-4e10-bf8c-32baee90c44b\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.511880 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host" (OuterVolumeSpecName: "host") pod "6f8824b5-d767-4e10-bf8c-32baee90c44b" (UID: "6f8824b5-d767-4e10-bf8c-32baee90c44b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.511955 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m4gd\" (UniqueName: \"kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd\") pod \"6f8824b5-d767-4e10-bf8c-32baee90c44b\" (UID: \"6f8824b5-d767-4e10-bf8c-32baee90c44b\") " Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.512478 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f8824b5-d767-4e10-bf8c-32baee90c44b-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.519348 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd" (OuterVolumeSpecName: "kube-api-access-4m4gd") pod "6f8824b5-d767-4e10-bf8c-32baee90c44b" (UID: "6f8824b5-d767-4e10-bf8c-32baee90c44b"). InnerVolumeSpecName "kube-api-access-4m4gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.616062 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m4gd\" (UniqueName: \"kubernetes.io/projected/6f8824b5-d767-4e10-bf8c-32baee90c44b-kube-api-access-4m4gd\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.633909 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3f1dbab-8a06-4b99-8f41-39179983cd6c/cinder-scheduler/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.697875 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3f1dbab-8a06-4b99-8f41-39179983cd6c/probe/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.730427 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-nnd2v"] Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.730462 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-nnd2v"] Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.849167 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a/cinder-volume/0.log" Oct 14 10:00:09 crc kubenswrapper[5002]: I1014 10:00:09.895069 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a/probe/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.040271 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8_a6cb9d07-96cd-42d8-91a1-230e4f6316f0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.092774 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d_34674ceb-4ae7-48fb-84be-afaae15ca5bf/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.282984 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/init/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.302625 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfafc96db089f239cb8eb1489d3c2b7cf9e336730082b3ac9bd73c9f7387aec3" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.302683 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-nnd2v" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.470921 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/init/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.665725 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/dnsmasq-dns/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.683465 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0b995566-e3f1-4c1c-a9e0-02874deed49b/glance-httpd/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.756574 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0b995566-e3f1-4c1c-a9e0-02874deed49b/glance-log/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.893548 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f64389b8-a21d-4cf5-abe8-402b170db4b3/glance-httpd/0.log" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.921075 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-fkqvg"] Oct 14 10:00:10 crc kubenswrapper[5002]: E1014 10:00:10.921507 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8824b5-d767-4e10-bf8c-32baee90c44b" containerName="container-00" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.921522 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8824b5-d767-4e10-bf8c-32baee90c44b" containerName="container-00" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.921796 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8824b5-d767-4e10-bf8c-32baee90c44b" containerName="container-00" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.922563 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:10 crc kubenswrapper[5002]: I1014 10:00:10.958280 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f64389b8-a21d-4cf5-abe8-402b170db4b3/glance-log/0.log" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.045651 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2gdh\" (UniqueName: \"kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.045825 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.146347 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-557d55dc9d-sk8l2_681029aa-9883-4641-9008-aeb673358cdd/horizon/0.log" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.148018 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2gdh\" (UniqueName: \"kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.148107 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.148241 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.168205 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2gdh\" (UniqueName: \"kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh\") pod \"crc-debug-fkqvg\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.239045 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:11 crc kubenswrapper[5002]: W1014 10:00:11.279421 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd961026c_7cb7_4e9e_adba_d7dda26a56ac.slice/crio-70f2d82fddd2c7d075a9dc0c3eb119bdfa9d15956bab4a12a0aaeee03ca90e9b WatchSource:0}: Error finding container 70f2d82fddd2c7d075a9dc0c3eb119bdfa9d15956bab4a12a0aaeee03ca90e9b: Status 404 returned error can't find the container with id 70f2d82fddd2c7d075a9dc0c3eb119bdfa9d15956bab4a12a0aaeee03ca90e9b Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.315203 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" event={"ID":"d961026c-7cb7-4e9e-adba-d7dda26a56ac","Type":"ContainerStarted","Data":"70f2d82fddd2c7d075a9dc0c3eb119bdfa9d15956bab4a12a0aaeee03ca90e9b"} Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.403153 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_df77ae37-bb4c-433c-bd13-bc48e1e6fa2c/horizontest-tests-horizontest/0.log" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.603120 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf_1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.732213 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8824b5-d767-4e10-bf8c-32baee90c44b" path="/var/lib/kubelet/pods/6f8824b5-d767-4e10-bf8c-32baee90c44b/volumes" Oct 14 10:00:11 crc kubenswrapper[5002]: I1014 10:00:11.798906 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sfkxq_f8e84847-a7b7-43a3-90c7-7a9c1ea92415/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.245637 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-557d55dc9d-sk8l2_681029aa-9883-4641-9008-aeb673358cdd/horizon-log/0.log" Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.324354 5002 generic.go:334] "Generic (PLEG): container finished" podID="d961026c-7cb7-4e9e-adba-d7dda26a56ac" containerID="e720a3f01e90abc436cf039ba1b0cb2c576bdc6ad6faf72e965659ef24d666f2" exitCode=0 Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.324399 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" event={"ID":"d961026c-7cb7-4e9e-adba-d7dda26a56ac","Type":"ContainerDied","Data":"e720a3f01e90abc436cf039ba1b0cb2c576bdc6ad6faf72e965659ef24d666f2"} Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.365644 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-fkqvg"] Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.373284 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pl9xf/crc-debug-fkqvg"] Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.511085 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29340541-twn6l_181a00bb-d0d0-483b-a51e-a84e308f06c8/keystone-cron/0.log" Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.556779 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_670693df-bbd2-4bf4-a705-8cc8a02d3d2f/kube-state-metrics/0.log" Oct 14 10:00:12 crc kubenswrapper[5002]: I1014 10:00:12.761626 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj_4d34c041-ecf1-4380-8fd9-6bd9a9660f87/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.008805 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_978c2b52-ead1-4ca3-924e-83ed4bb321a5/manila-api-log/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.121481 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_978c2b52-ead1-4ca3-924e-83ed4bb321a5/manila-api/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.338253 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_0bdd233c-7fae-4db7-8056-a218f6da8754/probe/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.343896 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_0bdd233c-7fae-4db7-8056-a218f6da8754/manila-scheduler/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.443266 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.555132 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d10fce14-aa5a-466f-948a-7787f2aa8a76/probe/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.597749 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d10fce14-aa5a-466f-948a-7787f2aa8a76/manila-share/0.log" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.602718 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host\") pod \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.602773 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host" (OuterVolumeSpecName: "host") pod "d961026c-7cb7-4e9e-adba-d7dda26a56ac" (UID: "d961026c-7cb7-4e9e-adba-d7dda26a56ac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.602792 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2gdh\" (UniqueName: \"kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh\") pod \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\" (UID: \"d961026c-7cb7-4e9e-adba-d7dda26a56ac\") " Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.603287 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d961026c-7cb7-4e9e-adba-d7dda26a56ac-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.608193 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh" (OuterVolumeSpecName: "kube-api-access-h2gdh") pod "d961026c-7cb7-4e9e-adba-d7dda26a56ac" (UID: "d961026c-7cb7-4e9e-adba-d7dda26a56ac"). InnerVolumeSpecName "kube-api-access-h2gdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.704501 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2gdh\" (UniqueName: \"kubernetes.io/projected/d961026c-7cb7-4e9e-adba-d7dda26a56ac-kube-api-access-h2gdh\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:13 crc kubenswrapper[5002]: I1014 10:00:13.730391 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d961026c-7cb7-4e9e-adba-d7dda26a56ac" path="/var/lib/kubelet/pods/d961026c-7cb7-4e9e-adba-d7dda26a56ac/volumes" Oct 14 10:00:14 crc kubenswrapper[5002]: I1014 10:00:14.343745 5002 scope.go:117] "RemoveContainer" containerID="e720a3f01e90abc436cf039ba1b0cb2c576bdc6ad6faf72e965659ef24d666f2" Oct 14 10:00:14 crc kubenswrapper[5002]: I1014 10:00:14.343807 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/crc-debug-fkqvg" Oct 14 10:00:14 crc kubenswrapper[5002]: I1014 10:00:14.457418 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7db98f685f-fphzv_e587d08e-64fa-4e35-98dc-ee283ada8c7f/keystone-api/0.log" Oct 14 10:00:14 crc kubenswrapper[5002]: I1014 10:00:14.797954 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b_721fe1de-3687-4a8c-babb-e30f9ec46abe/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:14 crc kubenswrapper[5002]: I1014 10:00:14.913329 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7855f7b865-dtwlh_abe829e6-d42f-447f-bcb2-aad97c1e4f92/neutron-httpd/0.log" Oct 14 10:00:15 crc kubenswrapper[5002]: I1014 10:00:15.347377 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7855f7b865-dtwlh_abe829e6-d42f-447f-bcb2-aad97c1e4f92/neutron-api/0.log" Oct 14 10:00:16 crc kubenswrapper[5002]: I1014 10:00:16.257020 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fc8e7450-29ac-4acb-a090-b7b0d1f2666f/nova-cell0-conductor-conductor/0.log" Oct 14 10:00:16 crc kubenswrapper[5002]: I1014 10:00:16.806643 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ba1b8fb3-2b74-47f6-adfe-58a6084311fb/nova-cell1-conductor-conductor/0.log" Oct 14 10:00:17 crc kubenswrapper[5002]: I1014 10:00:17.462375 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9b0062d-6204-46b8-8614-26f00ff2efd7/nova-cell1-novncproxy-novncproxy/0.log" Oct 14 10:00:17 crc kubenswrapper[5002]: I1014 10:00:17.936987 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5_99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:18 crc kubenswrapper[5002]: I1014 10:00:18.285961 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55d153d2-2f4c-4e77-9a7a-fa120ac94da2/nova-api-log/0.log" Oct 14 10:00:18 crc kubenswrapper[5002]: I1014 10:00:18.469306 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5/nova-metadata-log/0.log" Oct 14 10:00:18 crc kubenswrapper[5002]: I1014 10:00:18.720178 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:00:18 crc kubenswrapper[5002]: E1014 10:00:18.723937 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:00:19 crc kubenswrapper[5002]: I1014 10:00:19.535155 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_59facc52-a7a5-4a65-b7ba-67e9d8267de2/nova-scheduler-scheduler/0.log" Oct 14 10:00:19 crc kubenswrapper[5002]: I1014 10:00:19.605698 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55d153d2-2f4c-4e77-9a7a-fa120ac94da2/nova-api-api/0.log" Oct 14 10:00:19 crc kubenswrapper[5002]: I1014 10:00:19.809194 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/mysql-bootstrap/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.032319 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/galera/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.071295 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/mysql-bootstrap/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.287230 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/mysql-bootstrap/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.508661 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/mysql-bootstrap/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.566275 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/galera/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.766077 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c0df2486-02ae-48b5-b3b9-acb14ec03a6f/openstackclient/0.log" Oct 14 10:00:20 crc kubenswrapper[5002]: I1014 10:00:20.962546 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-b85t2_8622c6ea-480a-457b-b42c-2adad7acc425/openstack-network-exporter/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.250372 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server-init/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.448581 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server-init/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.489783 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovs-vswitchd/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.623006 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.804924 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tw5d7_f9439452-e42f-4fa8-ac61-65b5389ff828/ovn-controller/0.log" Oct 14 10:00:21 crc kubenswrapper[5002]: I1014 10:00:21.964686 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5/nova-metadata-metadata/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.030508 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-sz55g_0b20c844-7dba-4cdf-b6a2-b7deff3e401a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.209589 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3b8bf477-7229-46cd-aabc-2ae7794a694b/openstack-network-exporter/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.263852 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3b8bf477-7229-46cd-aabc-2ae7794a694b/ovn-northd/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.372159 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_705e9cf2-0759-49ed-8407-a73c3aed75cf/openstack-network-exporter/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.525088 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_705e9cf2-0759-49ed-8407-a73c3aed75cf/ovsdbserver-nb/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.721944 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2b959df9-bb11-4ee8-8af1-73beac51cfba/openstack-network-exporter/0.log" Oct 14 10:00:22 crc kubenswrapper[5002]: I1014 10:00:22.794523 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2b959df9-bb11-4ee8-8af1-73beac51cfba/ovsdbserver-sb/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.365544 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/setup-container/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.425723 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5dc947dd58-kwj5n_0141dd41-3526-47f8-8477-f0d347c58cea/placement-api/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.538286 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/setup-container/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.626334 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/rabbitmq/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.646236 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5dc947dd58-kwj5n_0141dd41-3526-47f8-8477-f0d347c58cea/placement-log/0.log" Oct 14 10:00:23 crc kubenswrapper[5002]: I1014 10:00:23.864594 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/setup-container/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.111983 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/setup-container/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.145370 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/rabbitmq/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.317991 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-f955z_3b28c2e3-928a-400a-92e3-c4e407257f94/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.433031 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2_fc36bcdf-48eb-4ef8-90c9-eb7842529652/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.597457 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hlm2l_75c3bce2-2453-4d20-a946-9a28fe2151e8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.733550 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-46bxg_b1704867-21c5-46a1-8d37-85af1337b303/ssh-known-hosts-edpm-deployment/0.log" Oct 14 10:00:24 crc kubenswrapper[5002]: I1014 10:00:24.904626 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_ec2597cf-678d-44a5-a3e3-41be976a3dc6/tempest-tests-tempest-tests-runner/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.133885 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_08748a2d-35ba-47e7-a01a-51f71b61d041/tempest-tests-tempest-tests-runner/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.223171 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_487e1aee-dbf2-451c-86dd-e2df3f229d89/test-operator-logs-container/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.265873 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d0a98c43-f790-476c-822c-e9b5957bbb78/memcached/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.297785 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1/test-operator-logs-container/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.431782 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c8166676-1bb1-4eec-9d75-28ccb7357ec3/test-operator-logs-container/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.536953 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_96b65a05-43d2-4d66-b467-3c88b9ce57e4/test-operator-logs-container/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.626802 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e/tobiko-tests-tobiko/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.766636 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_d8f34f80-d17b-4cdb-9834-f2e96de161ae/tobiko-tests-tobiko/0.log" Oct 14 10:00:25 crc kubenswrapper[5002]: I1014 10:00:25.848006 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn_7229bab8-aa3f-46dc-bcfc-338bce4801af/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:00:31 crc kubenswrapper[5002]: I1014 10:00:31.433649 5002 scope.go:117] "RemoveContainer" containerID="cb41609ad3f7fe3f33f33ef5f7e6b9122b9394414eddb5a7b69ceb9d88250bcd" Oct 14 10:00:31 crc kubenswrapper[5002]: I1014 10:00:31.720602 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:00:31 crc kubenswrapper[5002]: E1014 10:00:31.720859 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:00:45 crc kubenswrapper[5002]: I1014 10:00:45.720768 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:00:45 crc kubenswrapper[5002]: E1014 10:00:45.721954 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:00:48 crc kubenswrapper[5002]: I1014 10:00:48.996594 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.148170 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.189159 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.200391 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.343501 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.407242 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/extract/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.438527 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.670618 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-qpnq7_46630983-31be-46df-9e05-0506abbaf29e/kube-rbac-proxy/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.780642 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lr6gb_cfe90f21-7b6a-4364-8c33-92b2880e9e0b/kube-rbac-proxy/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.813481 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-qpnq7_46630983-31be-46df-9e05-0506abbaf29e/manager/0.log" Oct 14 10:00:49 crc kubenswrapper[5002]: I1014 10:00:49.911168 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lr6gb_cfe90f21-7b6a-4364-8c33-92b2880e9e0b/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.034645 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-9kfv7_578becbc-f063-4b98-b5a3-eb0656b24850/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.039468 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-9kfv7_578becbc-f063-4b98-b5a3-eb0656b24850/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.211493 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-8j2qn_de07a029-a377-47c8-be2f-a31148f4948a/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.242256 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-8j2qn_de07a029-a377-47c8-be2f-a31148f4948a/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.327099 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-5t4hm_63c73302-dc00-4dfe-9d3f-d80cb181ffc2/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.406341 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-5t4hm_63c73302-dc00-4dfe-9d3f-d80cb181ffc2/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.475056 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-s6vkc_bbc859e5-324a-413c-a7ac-499a0b0b8fba/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.563490 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-s6vkc_bbc859e5-324a-413c-a7ac-499a0b0b8fba/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.633387 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-d9ftz_17c37ba2-ff55-4b60-a1db-01dec0522e2e/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.807303 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-d9ftz_17c37ba2-ff55-4b60-a1db-01dec0522e2e/manager/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.862440 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-7jcbw_e1488db8-3d22-4c1f-8400-8d292b48027a/kube-rbac-proxy/0.log" Oct 14 10:00:50 crc kubenswrapper[5002]: I1014 10:00:50.903617 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-7jcbw_e1488db8-3d22-4c1f-8400-8d292b48027a/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.038529 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-kqdfb_4b405ac3-bbcd-4432-8f99-2862ef406541/kube-rbac-proxy/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.103097 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-kqdfb_4b405ac3-bbcd-4432-8f99-2862ef406541/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.276035 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-59jzf_2041ec73-97de-43ea-ae4e-e8a1d7521157/kube-rbac-proxy/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.338069 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-xsb2d_57a0f853-8feb-4064-9ed2-3460f2f66901/kube-rbac-proxy/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.344044 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-59jzf_2041ec73-97de-43ea-ae4e-e8a1d7521157/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.464223 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-xsb2d_57a0f853-8feb-4064-9ed2-3460f2f66901/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.550161 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-69rk7_95e654ed-72c1-4f5b-b670-a89c96692455/kube-rbac-proxy/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.555154 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-69rk7_95e654ed-72c1-4f5b-b670-a89c96692455/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.693970 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-82c6h_9c72f60f-c311-45eb-84a1-8c95164be376/kube-rbac-proxy/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.818277 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-82c6h_9c72f60f-c311-45eb-84a1-8c95164be376/manager/0.log" Oct 14 10:00:51 crc kubenswrapper[5002]: I1014 10:00:51.937352 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-cn7pq_27e01872-38da-4f89-8232-c8edaec917f8/kube-rbac-proxy/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.004743 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d44848n4xg8_05103074-be63-4240-bd7e-737c3c6ff4d5/kube-rbac-proxy/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.006465 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-cn7pq_27e01872-38da-4f89-8232-c8edaec917f8/manager/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.132433 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d44848n4xg8_05103074-be63-4240-bd7e-737c3c6ff4d5/manager/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.237300 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-k6scp_516be58e-c541-4fbe-b4b8-127e07a0b8b7/kube-rbac-proxy/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.404452 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd554b564-6djql_41ec772c-a692-4e45-90f1-5466203d5587/kube-rbac-proxy/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.502784 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd554b564-6djql_41ec772c-a692-4e45-90f1-5466203d5587/operator/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.800555 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-xj9vc_8e0847dd-54d1-44f7-8df9-b1378b00cb67/kube-rbac-proxy/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.863327 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pgk79_4f65b042-e91d-4c89-9aa4-52abc662f94f/registry-server/0.log" Oct 14 10:00:52 crc kubenswrapper[5002]: I1014 10:00:52.979136 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-xj9vc_8e0847dd-54d1-44f7-8df9-b1378b00cb67/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.021870 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-tws79_20fefb60-1453-4a27-9f7a-518df5b902fc/kube-rbac-proxy/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.061202 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-tws79_20fefb60-1453-4a27-9f7a-518df5b902fc/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.212649 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-s89dt_b1921718-5347-4bfd-a327-9c38671b4888/operator/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.303237 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-74c8t_b04b0d3d-03a0-4f12-814f-599f2899962b/kube-rbac-proxy/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.348246 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-k6scp_516be58e-c541-4fbe-b4b8-127e07a0b8b7/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.444615 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-74c8t_b04b0d3d-03a0-4f12-814f-599f2899962b/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.487054 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-fp4rw_42f32565-187e-4c71-a2c3-35eb988b8243/kube-rbac-proxy/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.566976 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-fp4rw_42f32565-187e-4c71-a2c3-35eb988b8243/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.639269 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c475b9f8d-pbb6v_cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f/kube-rbac-proxy/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.688675 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c475b9f8d-pbb6v_cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f/manager/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.764510 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-lwlwj_74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8/kube-rbac-proxy/0.log" Oct 14 10:00:53 crc kubenswrapper[5002]: I1014 10:00:53.783172 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-lwlwj_74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8/manager/0.log" Oct 14 10:00:57 crc kubenswrapper[5002]: I1014 10:00:57.751241 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:00:57 crc kubenswrapper[5002]: E1014 10:00:57.754341 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.162027 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29340601-5wkbp"] Oct 14 10:01:00 crc kubenswrapper[5002]: E1014 10:01:00.163455 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d961026c-7cb7-4e9e-adba-d7dda26a56ac" containerName="container-00" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.163496 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="d961026c-7cb7-4e9e-adba-d7dda26a56ac" containerName="container-00" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.163778 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="d961026c-7cb7-4e9e-adba-d7dda26a56ac" containerName="container-00" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.164791 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.170892 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340601-5wkbp"] Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.221429 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.221571 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtnc\" (UniqueName: \"kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.221670 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.221719 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.323459 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.323725 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.323845 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.323893 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtnc\" (UniqueName: \"kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.332027 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.332713 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.340345 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtnc\" (UniqueName: \"kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.348542 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle\") pod \"keystone-cron-29340601-5wkbp\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.492186 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:00 crc kubenswrapper[5002]: I1014 10:01:00.981926 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340601-5wkbp"] Oct 14 10:01:01 crc kubenswrapper[5002]: I1014 10:01:01.776281 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-5wkbp" event={"ID":"4d2fdadc-5e2c-4186-8402-229511267bb2","Type":"ContainerStarted","Data":"18a7ac6178008058510a616b4b359e7caa29a4fe68c2ff21e0a0a43750c7f9c8"} Oct 14 10:01:01 crc kubenswrapper[5002]: I1014 10:01:01.776355 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-5wkbp" event={"ID":"4d2fdadc-5e2c-4186-8402-229511267bb2","Type":"ContainerStarted","Data":"248c4e331df62083fe920222142705b67ad7b7f978dbccf19ef55ec9889aa521"} Oct 14 10:01:01 crc kubenswrapper[5002]: I1014 10:01:01.805508 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29340601-5wkbp" podStartSLOduration=1.8054887590000002 podStartE2EDuration="1.805488759s" podCreationTimestamp="2025-10-14 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:01:01.797445895 +0000 UTC m=+7794.778685367" watchObservedRunningTime="2025-10-14 10:01:01.805488759 +0000 UTC m=+7794.786728201" Oct 14 10:01:03 crc kubenswrapper[5002]: E1014 10:01:03.720172 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:01:04 crc kubenswrapper[5002]: I1014 10:01:04.805287 5002 generic.go:334] "Generic (PLEG): container finished" podID="4d2fdadc-5e2c-4186-8402-229511267bb2" containerID="18a7ac6178008058510a616b4b359e7caa29a4fe68c2ff21e0a0a43750c7f9c8" exitCode=0 Oct 14 10:01:04 crc kubenswrapper[5002]: I1014 10:01:04.805373 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-5wkbp" event={"ID":"4d2fdadc-5e2c-4186-8402-229511267bb2","Type":"ContainerDied","Data":"18a7ac6178008058510a616b4b359e7caa29a4fe68c2ff21e0a0a43750c7f9c8"} Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.208491 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.258147 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqtnc\" (UniqueName: \"kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc\") pod \"4d2fdadc-5e2c-4186-8402-229511267bb2\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.258428 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle\") pod \"4d2fdadc-5e2c-4186-8402-229511267bb2\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.258493 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys\") pod \"4d2fdadc-5e2c-4186-8402-229511267bb2\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.258534 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data\") pod \"4d2fdadc-5e2c-4186-8402-229511267bb2\" (UID: \"4d2fdadc-5e2c-4186-8402-229511267bb2\") " Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.266374 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4d2fdadc-5e2c-4186-8402-229511267bb2" (UID: "4d2fdadc-5e2c-4186-8402-229511267bb2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.266437 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc" (OuterVolumeSpecName: "kube-api-access-kqtnc") pod "4d2fdadc-5e2c-4186-8402-229511267bb2" (UID: "4d2fdadc-5e2c-4186-8402-229511267bb2"). InnerVolumeSpecName "kube-api-access-kqtnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.317411 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d2fdadc-5e2c-4186-8402-229511267bb2" (UID: "4d2fdadc-5e2c-4186-8402-229511267bb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.360795 5002 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.360829 5002 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.360851 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqtnc\" (UniqueName: \"kubernetes.io/projected/4d2fdadc-5e2c-4186-8402-229511267bb2-kube-api-access-kqtnc\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.386880 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data" (OuterVolumeSpecName: "config-data") pod "4d2fdadc-5e2c-4186-8402-229511267bb2" (UID: "4d2fdadc-5e2c-4186-8402-229511267bb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.462700 5002 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d2fdadc-5e2c-4186-8402-229511267bb2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.828294 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-5wkbp" event={"ID":"4d2fdadc-5e2c-4186-8402-229511267bb2","Type":"ContainerDied","Data":"248c4e331df62083fe920222142705b67ad7b7f978dbccf19ef55ec9889aa521"} Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.828337 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248c4e331df62083fe920222142705b67ad7b7f978dbccf19ef55ec9889aa521" Oct 14 10:01:06 crc kubenswrapper[5002]: I1014 10:01:06.828374 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-5wkbp" Oct 14 10:01:11 crc kubenswrapper[5002]: I1014 10:01:11.197957 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9gbqd_6b840956-6900-4c7c-89ce-622ddf455cf6/control-plane-machine-set-operator/0.log" Oct 14 10:01:11 crc kubenswrapper[5002]: I1014 10:01:11.387610 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hjt8h_71022b1b-51c8-4c8c-8d57-dc220ebf1b05/kube-rbac-proxy/0.log" Oct 14 10:01:11 crc kubenswrapper[5002]: I1014 10:01:11.394544 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hjt8h_71022b1b-51c8-4c8c-8d57-dc220ebf1b05/machine-api-operator/0.log" Oct 14 10:01:11 crc kubenswrapper[5002]: I1014 10:01:11.720630 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:01:11 crc kubenswrapper[5002]: E1014 10:01:11.720894 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:01:23 crc kubenswrapper[5002]: I1014 10:01:23.770891 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hsnk9_cbaa9318-0355-41df-b7bd-bd2fe2ce7d30/cert-manager-controller/0.log" Oct 14 10:01:23 crc kubenswrapper[5002]: I1014 10:01:23.959067 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-wkw2k_632fb195-4a92-40b0-96e7-4db2227146d5/cert-manager-cainjector/0.log" Oct 14 10:01:24 crc kubenswrapper[5002]: I1014 10:01:24.043791 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-66k4z_067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7/cert-manager-webhook/0.log" Oct 14 10:01:25 crc kubenswrapper[5002]: I1014 10:01:25.720446 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:01:25 crc kubenswrapper[5002]: E1014 10:01:25.721288 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.475933 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-s9gvm_59bfeb19-1a15-4cbd-8970-5c18ff149992/nmstate-console-plugin/0.log" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.648402 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-82rj9_0629a9d2-3340-4e00-9673-baaaf354a8ce/nmstate-handler/0.log" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.726475 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dq7k8_89f2e467-845f-45fb-b411-48c0cf2d3f0d/nmstate-metrics/0.log" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.776865 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dq7k8_89f2e467-845f-45fb-b411-48c0cf2d3f0d/kube-rbac-proxy/0.log" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.890530 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hpwzt_38ea0de3-d753-486f-8fc0-b5239e804cb2/nmstate-operator/0.log" Oct 14 10:01:36 crc kubenswrapper[5002]: I1014 10:01:36.956412 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-4tmr2_cb6927e5-34dd-43d1-9dea-db00339df27e/nmstate-webhook/0.log" Oct 14 10:01:37 crc kubenswrapper[5002]: I1014 10:01:37.732813 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:01:37 crc kubenswrapper[5002]: E1014 10:01:37.733102 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:01:49 crc kubenswrapper[5002]: I1014 10:01:49.721024 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:01:49 crc kubenswrapper[5002]: E1014 10:01:49.722035 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:01:51 crc kubenswrapper[5002]: I1014 10:01:51.637672 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2zwcb_0a00fbf0-fcf8-42f8-a8da-cbf266655ee3/kube-rbac-proxy/0.log" Oct 14 10:01:51 crc kubenswrapper[5002]: I1014 10:01:51.788540 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2zwcb_0a00fbf0-fcf8-42f8-a8da-cbf266655ee3/controller/0.log" Oct 14 10:01:51 crc kubenswrapper[5002]: I1014 10:01:51.883658 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.023341 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.023467 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.053526 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.057946 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.240687 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.258472 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.271955 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.280588 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.448448 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.449698 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/controller/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.478332 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.480872 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.616981 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/frr-metrics/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.686297 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/kube-rbac-proxy/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.730298 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/kube-rbac-proxy-frr/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.916627 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/reloader/0.log" Oct 14 10:01:52 crc kubenswrapper[5002]: I1014 10:01:52.984261 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-lf7b4_10d5d438-0656-4699-b386-a539ca6810d2/frr-k8s-webhook-server/0.log" Oct 14 10:01:53 crc kubenswrapper[5002]: I1014 10:01:53.136113 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7769bf444d-r6x5z_45cb6486-ddee-4491-86d3-6de31fa2eeed/manager/0.log" Oct 14 10:01:53 crc kubenswrapper[5002]: I1014 10:01:53.384394 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74b78f565b-6dh62_513044aa-d565-47b2-8fb7-5df37788a795/webhook-server/0.log" Oct 14 10:01:53 crc kubenswrapper[5002]: I1014 10:01:53.411917 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-686b6_6d62cfe7-6906-4f1f-bc0b-595c98d4af86/kube-rbac-proxy/0.log" Oct 14 10:01:53 crc kubenswrapper[5002]: I1014 10:01:53.997293 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-686b6_6d62cfe7-6906-4f1f-bc0b-595c98d4af86/speaker/0.log" Oct 14 10:01:54 crc kubenswrapper[5002]: I1014 10:01:54.579919 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/frr/0.log" Oct 14 10:02:04 crc kubenswrapper[5002]: I1014 10:02:04.721246 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:02:04 crc kubenswrapper[5002]: E1014 10:02:04.722060 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.318831 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.497055 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.498351 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.506128 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.639548 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.653527 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.660702 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/extract/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.793352 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.940945 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.944346 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:02:07 crc kubenswrapper[5002]: I1014 10:02:07.972974 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.105948 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.133402 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.314120 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.538964 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.574478 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.584304 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.773730 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:02:08 crc kubenswrapper[5002]: I1014 10:02:08.848173 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.103108 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.107489 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/registry-server/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.244167 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.342117 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.354039 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.454438 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:09 crc kubenswrapper[5002]: E1014 10:02:09.454854 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2fdadc-5e2c-4186-8402-229511267bb2" containerName="keystone-cron" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.454891 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2fdadc-5e2c-4186-8402-229511267bb2" containerName="keystone-cron" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.455103 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d2fdadc-5e2c-4186-8402-229511267bb2" containerName="keystone-cron" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.456684 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.516554 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.526703 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.527456 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f66j\" (UniqueName: \"kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.527626 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.629401 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.629554 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.629601 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f66j\" (UniqueName: \"kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.629915 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.630000 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.663662 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f66j\" (UniqueName: \"kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j\") pod \"community-operators-wzsj9\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.687684 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.755233 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.768513 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/extract/0.log" Oct 14 10:02:09 crc kubenswrapper[5002]: I1014 10:02:09.787709 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.033266 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/registry-server/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.237511 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cq8mp_651e3f67-0cad-4654-8c98-0672dd56f367/marketplace-operator/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.330856 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.346332 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.499483 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerStarted","Data":"705878c9e426098ccd5e4dc9a0f79ccba1e0edaab8a78778c9d52d95c4b48c47"} Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.554156 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.564212 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.567235 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.714613 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.716467 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:02:10 crc kubenswrapper[5002]: I1014 10:02:10.906986 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.032208 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/registry-server/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.064617 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.066413 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.078851 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.281364 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.301274 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.508599 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerID="546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5" exitCode=0 Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.508911 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerDied","Data":"546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5"} Oct 14 10:02:11 crc kubenswrapper[5002]: I1014 10:02:11.510923 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:02:12 crc kubenswrapper[5002]: I1014 10:02:12.273309 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/registry-server/0.log" Oct 14 10:02:12 crc kubenswrapper[5002]: I1014 10:02:12.520402 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerStarted","Data":"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73"} Oct 14 10:02:13 crc kubenswrapper[5002]: I1014 10:02:13.530976 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerID="4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73" exitCode=0 Oct 14 10:02:13 crc kubenswrapper[5002]: I1014 10:02:13.531025 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerDied","Data":"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73"} Oct 14 10:02:14 crc kubenswrapper[5002]: I1014 10:02:14.547695 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerStarted","Data":"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d"} Oct 14 10:02:14 crc kubenswrapper[5002]: I1014 10:02:14.571964 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wzsj9" podStartSLOduration=2.992451376 podStartE2EDuration="5.57193258s" podCreationTimestamp="2025-10-14 10:02:09 +0000 UTC" firstStartedPulling="2025-10-14 10:02:11.51067174 +0000 UTC m=+7864.491911192" lastFinishedPulling="2025-10-14 10:02:14.090152914 +0000 UTC m=+7867.071392396" observedRunningTime="2025-10-14 10:02:14.568442247 +0000 UTC m=+7867.549681729" watchObservedRunningTime="2025-10-14 10:02:14.57193258 +0000 UTC m=+7867.553172032" Oct 14 10:02:19 crc kubenswrapper[5002]: I1014 10:02:19.720823 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:02:19 crc kubenswrapper[5002]: E1014 10:02:19.721606 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:02:19 crc kubenswrapper[5002]: I1014 10:02:19.788692 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:19 crc kubenswrapper[5002]: I1014 10:02:19.788729 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:19 crc kubenswrapper[5002]: I1014 10:02:19.866555 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:20 crc kubenswrapper[5002]: I1014 10:02:20.676374 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:20 crc kubenswrapper[5002]: I1014 10:02:20.734636 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:22 crc kubenswrapper[5002]: I1014 10:02:22.637149 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wzsj9" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="registry-server" containerID="cri-o://b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d" gracePeriod=2 Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.150215 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.222702 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content\") pod \"f4d83163-360c-4393-9f6a-91ecbc399ee4\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.223012 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f66j\" (UniqueName: \"kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j\") pod \"f4d83163-360c-4393-9f6a-91ecbc399ee4\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.223100 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities\") pod \"f4d83163-360c-4393-9f6a-91ecbc399ee4\" (UID: \"f4d83163-360c-4393-9f6a-91ecbc399ee4\") " Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.224010 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities" (OuterVolumeSpecName: "utilities") pod "f4d83163-360c-4393-9f6a-91ecbc399ee4" (UID: "f4d83163-360c-4393-9f6a-91ecbc399ee4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.237497 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j" (OuterVolumeSpecName: "kube-api-access-5f66j") pod "f4d83163-360c-4393-9f6a-91ecbc399ee4" (UID: "f4d83163-360c-4393-9f6a-91ecbc399ee4"). InnerVolumeSpecName "kube-api-access-5f66j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.272403 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4d83163-360c-4393-9f6a-91ecbc399ee4" (UID: "f4d83163-360c-4393-9f6a-91ecbc399ee4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.325540 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.325572 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d83163-360c-4393-9f6a-91ecbc399ee4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.325586 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f66j\" (UniqueName: \"kubernetes.io/projected/f4d83163-360c-4393-9f6a-91ecbc399ee4-kube-api-access-5f66j\") on node \"crc\" DevicePath \"\"" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.646809 5002 generic.go:334] "Generic (PLEG): container finished" podID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerID="b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d" exitCode=0 Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.646906 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzsj9" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.646930 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerDied","Data":"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d"} Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.648046 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzsj9" event={"ID":"f4d83163-360c-4393-9f6a-91ecbc399ee4","Type":"ContainerDied","Data":"705878c9e426098ccd5e4dc9a0f79ccba1e0edaab8a78778c9d52d95c4b48c47"} Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.648070 5002 scope.go:117] "RemoveContainer" containerID="b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.671390 5002 scope.go:117] "RemoveContainer" containerID="4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.695722 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.705761 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wzsj9"] Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.709837 5002 scope.go:117] "RemoveContainer" containerID="546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.740911 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" path="/var/lib/kubelet/pods/f4d83163-360c-4393-9f6a-91ecbc399ee4/volumes" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.746223 5002 scope.go:117] "RemoveContainer" containerID="b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d" Oct 14 10:02:23 crc kubenswrapper[5002]: E1014 10:02:23.746646 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d\": container with ID starting with b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d not found: ID does not exist" containerID="b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.746682 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d"} err="failed to get container status \"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d\": rpc error: code = NotFound desc = could not find container \"b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d\": container with ID starting with b7c94c656ea27d545568da4f147d760b0fb1c3ab917c71d398ae54adec2ac85d not found: ID does not exist" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.746711 5002 scope.go:117] "RemoveContainer" containerID="4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73" Oct 14 10:02:23 crc kubenswrapper[5002]: E1014 10:02:23.746972 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73\": container with ID starting with 4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73 not found: ID does not exist" containerID="4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.746987 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73"} err="failed to get container status \"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73\": rpc error: code = NotFound desc = could not find container \"4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73\": container with ID starting with 4721da66b0887f660408b7f32f6a18724d86f2c21bea86040812508397bd0d73 not found: ID does not exist" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.746999 5002 scope.go:117] "RemoveContainer" containerID="546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5" Oct 14 10:02:23 crc kubenswrapper[5002]: E1014 10:02:23.747195 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5\": container with ID starting with 546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5 not found: ID does not exist" containerID="546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5" Oct 14 10:02:23 crc kubenswrapper[5002]: I1014 10:02:23.747214 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5"} err="failed to get container status \"546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5\": rpc error: code = NotFound desc = could not find container \"546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5\": container with ID starting with 546267c764f180b00cacf366e62e4fff01b75d91f1ffdfdcc06a6bd0ffe959e5 not found: ID does not exist" Oct 14 10:02:29 crc kubenswrapper[5002]: E1014 10:02:29.720677 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:02:32 crc kubenswrapper[5002]: I1014 10:02:32.720251 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:02:32 crc kubenswrapper[5002]: E1014 10:02:32.721124 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:02:47 crc kubenswrapper[5002]: I1014 10:02:47.748359 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:02:47 crc kubenswrapper[5002]: E1014 10:02:47.749145 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:03:01 crc kubenswrapper[5002]: I1014 10:03:01.721365 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:03:01 crc kubenswrapper[5002]: E1014 10:03:01.722407 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:03:13 crc kubenswrapper[5002]: I1014 10:03:13.721934 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:03:13 crc kubenswrapper[5002]: E1014 10:03:13.722641 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:03:25 crc kubenswrapper[5002]: I1014 10:03:25.722184 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:03:25 crc kubenswrapper[5002]: E1014 10:03:25.723694 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:03:38 crc kubenswrapper[5002]: I1014 10:03:38.720948 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:03:38 crc kubenswrapper[5002]: E1014 10:03:38.722122 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:03:53 crc kubenswrapper[5002]: I1014 10:03:53.720688 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:03:53 crc kubenswrapper[5002]: E1014 10:03:53.721649 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:03:53 crc kubenswrapper[5002]: E1014 10:03:53.721697 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:04:08 crc kubenswrapper[5002]: I1014 10:04:08.721329 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:04:08 crc kubenswrapper[5002]: E1014 10:04:08.722002 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:04:17 crc kubenswrapper[5002]: I1014 10:04:17.039643 5002 generic.go:334] "Generic (PLEG): container finished" podID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerID="b2e2073c1d98ef99ac374b68b0ae32b00728f6b231aceea1ac9d06dd3cbc5f9e" exitCode=0 Oct 14 10:04:17 crc kubenswrapper[5002]: I1014 10:04:17.039773 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" event={"ID":"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46","Type":"ContainerDied","Data":"b2e2073c1d98ef99ac374b68b0ae32b00728f6b231aceea1ac9d06dd3cbc5f9e"} Oct 14 10:04:17 crc kubenswrapper[5002]: I1014 10:04:17.040879 5002 scope.go:117] "RemoveContainer" containerID="b2e2073c1d98ef99ac374b68b0ae32b00728f6b231aceea1ac9d06dd3cbc5f9e" Oct 14 10:04:17 crc kubenswrapper[5002]: I1014 10:04:17.561245 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pl9xf_must-gather-jsjrs_9664c77f-9e74-4ce1-ba2a-0f03c23e8b46/gather/0.log" Oct 14 10:04:20 crc kubenswrapper[5002]: I1014 10:04:20.720094 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:04:20 crc kubenswrapper[5002]: E1014 10:04:20.720921 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:04:25 crc kubenswrapper[5002]: I1014 10:04:25.916151 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pl9xf/must-gather-jsjrs"] Oct 14 10:04:25 crc kubenswrapper[5002]: I1014 10:04:25.916990 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="copy" containerID="cri-o://efa175bf4460d11990a9d5970622695b7b24c7e79cea897d117f0503b90e5d9b" gracePeriod=2 Oct 14 10:04:25 crc kubenswrapper[5002]: I1014 10:04:25.925248 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pl9xf/must-gather-jsjrs"] Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.161730 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pl9xf_must-gather-jsjrs_9664c77f-9e74-4ce1-ba2a-0f03c23e8b46/copy/0.log" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.162240 5002 generic.go:334] "Generic (PLEG): container finished" podID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerID="efa175bf4460d11990a9d5970622695b7b24c7e79cea897d117f0503b90e5d9b" exitCode=143 Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.365517 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pl9xf_must-gather-jsjrs_9664c77f-9e74-4ce1-ba2a-0f03c23e8b46/copy/0.log" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.366375 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.468272 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output\") pod \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.468540 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gxs4\" (UniqueName: \"kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4\") pod \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\" (UID: \"9664c77f-9e74-4ce1-ba2a-0f03c23e8b46\") " Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.474384 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4" (OuterVolumeSpecName: "kube-api-access-2gxs4") pod "9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" (UID: "9664c77f-9e74-4ce1-ba2a-0f03c23e8b46"). InnerVolumeSpecName "kube-api-access-2gxs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.571520 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gxs4\" (UniqueName: \"kubernetes.io/projected/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-kube-api-access-2gxs4\") on node \"crc\" DevicePath \"\"" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.658249 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" (UID: "9664c77f-9e74-4ce1-ba2a-0f03c23e8b46"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:04:26 crc kubenswrapper[5002]: I1014 10:04:26.676185 5002 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 10:04:27 crc kubenswrapper[5002]: I1014 10:04:27.172271 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pl9xf_must-gather-jsjrs_9664c77f-9e74-4ce1-ba2a-0f03c23e8b46/copy/0.log" Oct 14 10:04:27 crc kubenswrapper[5002]: I1014 10:04:27.172760 5002 scope.go:117] "RemoveContainer" containerID="efa175bf4460d11990a9d5970622695b7b24c7e79cea897d117f0503b90e5d9b" Oct 14 10:04:27 crc kubenswrapper[5002]: I1014 10:04:27.172803 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pl9xf/must-gather-jsjrs" Oct 14 10:04:27 crc kubenswrapper[5002]: I1014 10:04:27.203507 5002 scope.go:117] "RemoveContainer" containerID="b2e2073c1d98ef99ac374b68b0ae32b00728f6b231aceea1ac9d06dd3cbc5f9e" Oct 14 10:04:27 crc kubenswrapper[5002]: I1014 10:04:27.729891 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" path="/var/lib/kubelet/pods/9664c77f-9e74-4ce1-ba2a-0f03c23e8b46/volumes" Oct 14 10:04:31 crc kubenswrapper[5002]: I1014 10:04:31.721410 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:04:31 crc kubenswrapper[5002]: E1014 10:04:31.722791 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:04:44 crc kubenswrapper[5002]: I1014 10:04:44.720594 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:04:45 crc kubenswrapper[5002]: I1014 10:04:45.390761 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb"} Oct 14 10:04:56 crc kubenswrapper[5002]: E1014 10:04:56.720401 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.726262 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k2pt5/must-gather-2dtds"] Oct 14 10:05:24 crc kubenswrapper[5002]: E1014 10:05:24.727925 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="registry-server" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728032 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="registry-server" Oct 14 10:05:24 crc kubenswrapper[5002]: E1014 10:05:24.728122 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="gather" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728184 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="gather" Oct 14 10:05:24 crc kubenswrapper[5002]: E1014 10:05:24.728250 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="copy" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728306 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="copy" Oct 14 10:05:24 crc kubenswrapper[5002]: E1014 10:05:24.728385 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="extract-content" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728446 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="extract-content" Oct 14 10:05:24 crc kubenswrapper[5002]: E1014 10:05:24.728515 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="extract-utilities" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728571 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="extract-utilities" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728821 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="copy" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728918 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="9664c77f-9e74-4ce1-ba2a-0f03c23e8b46" containerName="gather" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.728990 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d83163-360c-4393-9f6a-91ecbc399ee4" containerName="registry-server" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.729978 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.736369 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k2pt5"/"kube-root-ca.crt" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.736911 5002 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-k2pt5"/"openshift-service-ca.crt" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.737347 5002 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-k2pt5"/"default-dockercfg-j22r2" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.748223 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k2pt5/must-gather-2dtds"] Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.761199 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.761356 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdxj\" (UniqueName: \"kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.863906 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.864059 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdxj\" (UniqueName: \"kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.864447 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:24 crc kubenswrapper[5002]: I1014 10:05:24.882409 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdxj\" (UniqueName: \"kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj\") pod \"must-gather-2dtds\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:25 crc kubenswrapper[5002]: I1014 10:05:25.051481 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:05:25 crc kubenswrapper[5002]: I1014 10:05:25.515999 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-k2pt5/must-gather-2dtds"] Oct 14 10:05:25 crc kubenswrapper[5002]: I1014 10:05:25.862337 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/must-gather-2dtds" event={"ID":"3eb989ee-6138-47c7-8d09-dedcf1364fd2","Type":"ContainerStarted","Data":"718a01b93b7cb686b6d41dff767e8cd66e65b6de0efefc85f5c1a7267467f730"} Oct 14 10:05:25 crc kubenswrapper[5002]: I1014 10:05:25.862706 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/must-gather-2dtds" event={"ID":"3eb989ee-6138-47c7-8d09-dedcf1364fd2","Type":"ContainerStarted","Data":"bc41322aa2c44c4d19f4ff290a85c567ac44f54bc682c95c516398327c60e441"} Oct 14 10:05:26 crc kubenswrapper[5002]: I1014 10:05:26.874382 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/must-gather-2dtds" event={"ID":"3eb989ee-6138-47c7-8d09-dedcf1364fd2","Type":"ContainerStarted","Data":"f346cc4a1dc60de31a83a0d527785353d09b17b196d072ffb455d987700d1f54"} Oct 14 10:05:26 crc kubenswrapper[5002]: I1014 10:05:26.902641 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k2pt5/must-gather-2dtds" podStartSLOduration=2.902613676 podStartE2EDuration="2.902613676s" podCreationTimestamp="2025-10-14 10:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:05:26.890201465 +0000 UTC m=+8059.871440947" watchObservedRunningTime="2025-10-14 10:05:26.902613676 +0000 UTC m=+8059.883853158" Oct 14 10:05:29 crc kubenswrapper[5002]: I1014 10:05:29.990026 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-mctvg"] Oct 14 10:05:29 crc kubenswrapper[5002]: I1014 10:05:29.991770 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.079926 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t66d\" (UniqueName: \"kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.080304 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.182170 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t66d\" (UniqueName: \"kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.182293 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.182394 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.209498 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t66d\" (UniqueName: \"kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d\") pod \"crc-debug-mctvg\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.310520 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:05:30 crc kubenswrapper[5002]: W1014 10:05:30.336604 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0a48380_1433_49df_afca_e3f47d743e33.slice/crio-1cb82f2775dce329fce1a3d9f247eea3873455639578d26b90b02b791684eb64 WatchSource:0}: Error finding container 1cb82f2775dce329fce1a3d9f247eea3873455639578d26b90b02b791684eb64: Status 404 returned error can't find the container with id 1cb82f2775dce329fce1a3d9f247eea3873455639578d26b90b02b791684eb64 Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.907604 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" event={"ID":"f0a48380-1433-49df-afca-e3f47d743e33","Type":"ContainerStarted","Data":"ef86cc8037c152771a507a4359db777da077d00262561af7c800284b74173204"} Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.907970 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" event={"ID":"f0a48380-1433-49df-afca-e3f47d743e33","Type":"ContainerStarted","Data":"1cb82f2775dce329fce1a3d9f247eea3873455639578d26b90b02b791684eb64"} Oct 14 10:05:30 crc kubenswrapper[5002]: I1014 10:05:30.938329 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" podStartSLOduration=1.938307496 podStartE2EDuration="1.938307496s" podCreationTimestamp="2025-10-14 10:05:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:05:30.932511272 +0000 UTC m=+8063.913750754" watchObservedRunningTime="2025-10-14 10:05:30.938307496 +0000 UTC m=+8063.919546948" Oct 14 10:05:58 crc kubenswrapper[5002]: E1014 10:05:58.720683 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:06:11 crc kubenswrapper[5002]: I1014 10:06:11.311235 5002 generic.go:334] "Generic (PLEG): container finished" podID="f0a48380-1433-49df-afca-e3f47d743e33" containerID="ef86cc8037c152771a507a4359db777da077d00262561af7c800284b74173204" exitCode=0 Oct 14 10:06:11 crc kubenswrapper[5002]: I1014 10:06:11.311358 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" event={"ID":"f0a48380-1433-49df-afca-e3f47d743e33","Type":"ContainerDied","Data":"ef86cc8037c152771a507a4359db777da077d00262561af7c800284b74173204"} Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.436284 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.487810 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-mctvg"] Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.497380 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-mctvg"] Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.565642 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t66d\" (UniqueName: \"kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d\") pod \"f0a48380-1433-49df-afca-e3f47d743e33\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.565711 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host\") pod \"f0a48380-1433-49df-afca-e3f47d743e33\" (UID: \"f0a48380-1433-49df-afca-e3f47d743e33\") " Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.565781 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host" (OuterVolumeSpecName: "host") pod "f0a48380-1433-49df-afca-e3f47d743e33" (UID: "f0a48380-1433-49df-afca-e3f47d743e33"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.566395 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f0a48380-1433-49df-afca-e3f47d743e33-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.572130 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d" (OuterVolumeSpecName: "kube-api-access-7t66d") pod "f0a48380-1433-49df-afca-e3f47d743e33" (UID: "f0a48380-1433-49df-afca-e3f47d743e33"). InnerVolumeSpecName "kube-api-access-7t66d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:06:12 crc kubenswrapper[5002]: I1014 10:06:12.668195 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t66d\" (UniqueName: \"kubernetes.io/projected/f0a48380-1433-49df-afca-e3f47d743e33-kube-api-access-7t66d\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.347376 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cb82f2775dce329fce1a3d9f247eea3873455639578d26b90b02b791684eb64" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.347477 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-mctvg" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.652884 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-zjqdx"] Oct 14 10:06:13 crc kubenswrapper[5002]: E1014 10:06:13.653264 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a48380-1433-49df-afca-e3f47d743e33" containerName="container-00" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.653278 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a48380-1433-49df-afca-e3f47d743e33" containerName="container-00" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.653524 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a48380-1433-49df-afca-e3f47d743e33" containerName="container-00" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.654131 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.731743 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a48380-1433-49df-afca-e3f47d743e33" path="/var/lib/kubelet/pods/f0a48380-1433-49df-afca-e3f47d743e33/volumes" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.795087 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.795192 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzgdh\" (UniqueName: \"kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.896914 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.897295 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzgdh\" (UniqueName: \"kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.897078 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.916771 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzgdh\" (UniqueName: \"kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh\") pod \"crc-debug-zjqdx\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:13 crc kubenswrapper[5002]: I1014 10:06:13.978337 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:14 crc kubenswrapper[5002]: I1014 10:06:14.357764 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" event={"ID":"10c338a8-afb8-4e06-b70e-58b762a71ffc","Type":"ContainerStarted","Data":"8e6a7567574c02b886793172a9b80670de09b56dd581de474962d7bd184a8480"} Oct 14 10:06:14 crc kubenswrapper[5002]: I1014 10:06:14.358132 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" event={"ID":"10c338a8-afb8-4e06-b70e-58b762a71ffc","Type":"ContainerStarted","Data":"ed6576f830053e885cb3925660a2481320ac8e8f91a0f8dc21a6b00c38a62023"} Oct 14 10:06:14 crc kubenswrapper[5002]: I1014 10:06:14.379699 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" podStartSLOduration=1.3796785360000001 podStartE2EDuration="1.379678536s" podCreationTimestamp="2025-10-14 10:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:06:14.37347276 +0000 UTC m=+8107.354712232" watchObservedRunningTime="2025-10-14 10:06:14.379678536 +0000 UTC m=+8107.360917998" Oct 14 10:06:15 crc kubenswrapper[5002]: I1014 10:06:15.366235 5002 generic.go:334] "Generic (PLEG): container finished" podID="10c338a8-afb8-4e06-b70e-58b762a71ffc" containerID="8e6a7567574c02b886793172a9b80670de09b56dd581de474962d7bd184a8480" exitCode=0 Oct 14 10:06:15 crc kubenswrapper[5002]: I1014 10:06:15.366273 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" event={"ID":"10c338a8-afb8-4e06-b70e-58b762a71ffc","Type":"ContainerDied","Data":"8e6a7567574c02b886793172a9b80670de09b56dd581de474962d7bd184a8480"} Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.488152 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.641515 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzgdh\" (UniqueName: \"kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh\") pod \"10c338a8-afb8-4e06-b70e-58b762a71ffc\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.641557 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host\") pod \"10c338a8-afb8-4e06-b70e-58b762a71ffc\" (UID: \"10c338a8-afb8-4e06-b70e-58b762a71ffc\") " Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.641720 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host" (OuterVolumeSpecName: "host") pod "10c338a8-afb8-4e06-b70e-58b762a71ffc" (UID: "10c338a8-afb8-4e06-b70e-58b762a71ffc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.642242 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10c338a8-afb8-4e06-b70e-58b762a71ffc-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.653154 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh" (OuterVolumeSpecName: "kube-api-access-hzgdh") pod "10c338a8-afb8-4e06-b70e-58b762a71ffc" (UID: "10c338a8-afb8-4e06-b70e-58b762a71ffc"). InnerVolumeSpecName "kube-api-access-hzgdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.743655 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzgdh\" (UniqueName: \"kubernetes.io/projected/10c338a8-afb8-4e06-b70e-58b762a71ffc-kube-api-access-hzgdh\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.871751 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-zjqdx"] Oct 14 10:06:16 crc kubenswrapper[5002]: I1014 10:06:16.883089 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-zjqdx"] Oct 14 10:06:17 crc kubenswrapper[5002]: I1014 10:06:17.384279 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed6576f830053e885cb3925660a2481320ac8e8f91a0f8dc21a6b00c38a62023" Oct 14 10:06:17 crc kubenswrapper[5002]: I1014 10:06:17.384380 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-zjqdx" Oct 14 10:06:17 crc kubenswrapper[5002]: I1014 10:06:17.737317 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c338a8-afb8-4e06-b70e-58b762a71ffc" path="/var/lib/kubelet/pods/10c338a8-afb8-4e06-b70e-58b762a71ffc/volumes" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.028621 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-7j4sq"] Oct 14 10:06:18 crc kubenswrapper[5002]: E1014 10:06:18.028984 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c338a8-afb8-4e06-b70e-58b762a71ffc" containerName="container-00" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.029000 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c338a8-afb8-4e06-b70e-58b762a71ffc" containerName="container-00" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.029206 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c338a8-afb8-4e06-b70e-58b762a71ffc" containerName="container-00" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.029762 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.170577 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncrfg\" (UniqueName: \"kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.170696 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.274536 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncrfg\" (UniqueName: \"kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.274633 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.274948 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.289990 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncrfg\" (UniqueName: \"kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg\") pod \"crc-debug-7j4sq\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.355822 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:18 crc kubenswrapper[5002]: W1014 10:06:18.383201 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcc47e8d_d884_4ecc_a1ee_c0acf7c12f00.slice/crio-86ea3e756249112b37f925943e4e0f87784dd97495ddc9b707ce4bae652b2676 WatchSource:0}: Error finding container 86ea3e756249112b37f925943e4e0f87784dd97495ddc9b707ce4bae652b2676: Status 404 returned error can't find the container with id 86ea3e756249112b37f925943e4e0f87784dd97495ddc9b707ce4bae652b2676 Oct 14 10:06:18 crc kubenswrapper[5002]: I1014 10:06:18.408339 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" event={"ID":"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00","Type":"ContainerStarted","Data":"86ea3e756249112b37f925943e4e0f87784dd97495ddc9b707ce4bae652b2676"} Oct 14 10:06:19 crc kubenswrapper[5002]: I1014 10:06:19.417858 5002 generic.go:334] "Generic (PLEG): container finished" podID="bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" containerID="a19daf39bbe347d346a1c516a2c1a68a5cad8e6d0b9db2e8f2b742bbceab194a" exitCode=0 Oct 14 10:06:19 crc kubenswrapper[5002]: I1014 10:06:19.419202 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" event={"ID":"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00","Type":"ContainerDied","Data":"a19daf39bbe347d346a1c516a2c1a68a5cad8e6d0b9db2e8f2b742bbceab194a"} Oct 14 10:06:19 crc kubenswrapper[5002]: I1014 10:06:19.460756 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-7j4sq"] Oct 14 10:06:19 crc kubenswrapper[5002]: I1014 10:06:19.470243 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k2pt5/crc-debug-7j4sq"] Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.523637 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.619350 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host\") pod \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.619478 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host" (OuterVolumeSpecName: "host") pod "bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" (UID: "bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.619680 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncrfg\" (UniqueName: \"kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg\") pod \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\" (UID: \"bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00\") " Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.620266 5002 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.627034 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg" (OuterVolumeSpecName: "kube-api-access-ncrfg") pod "bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" (UID: "bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00"). InnerVolumeSpecName "kube-api-access-ncrfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:06:20 crc kubenswrapper[5002]: I1014 10:06:20.721722 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncrfg\" (UniqueName: \"kubernetes.io/projected/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00-kube-api-access-ncrfg\") on node \"crc\" DevicePath \"\"" Oct 14 10:06:21 crc kubenswrapper[5002]: I1014 10:06:21.438985 5002 scope.go:117] "RemoveContainer" containerID="a19daf39bbe347d346a1c516a2c1a68a5cad8e6d0b9db2e8f2b742bbceab194a" Oct 14 10:06:21 crc kubenswrapper[5002]: I1014 10:06:21.439032 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/crc-debug-7j4sq" Oct 14 10:06:21 crc kubenswrapper[5002]: I1014 10:06:21.730423 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" path="/var/lib/kubelet/pods/bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00/volumes" Oct 14 10:06:31 crc kubenswrapper[5002]: I1014 10:06:31.663122 5002 scope.go:117] "RemoveContainer" containerID="fd2ba8e7a6c433d9bb274d54a086342774d51273612a3a09527953a6b620c988" Oct 14 10:06:32 crc kubenswrapper[5002]: I1014 10:06:32.372096 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_0d845551-0fcc-4cc0-8d67-7dedb66eabb5/ansibletest-ansibletest/0.log" Oct 14 10:06:32 crc kubenswrapper[5002]: I1014 10:06:32.577124 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67d6789ffb-z5pvg_e3b04e6e-b1f7-4327-985d-6ed4e526eeaa/barbican-api-log/0.log" Oct 14 10:06:32 crc kubenswrapper[5002]: I1014 10:06:32.585259 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-67d6789ffb-z5pvg_e3b04e6e-b1f7-4327-985d-6ed4e526eeaa/barbican-api/0.log" Oct 14 10:06:32 crc kubenswrapper[5002]: I1014 10:06:32.764815 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f846844d6-pvmrs_a1f84982-1a7e-4054-b3f5-666de26a4e96/barbican-keystone-listener/0.log" Oct 14 10:06:32 crc kubenswrapper[5002]: I1014 10:06:32.964387 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7fffd76679-n928f_ec0b4c68-f07d-4d42-8225-6950e9315970/barbican-worker/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.213941 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7f846844d6-pvmrs_a1f84982-1a7e-4054-b3f5-666de26a4e96/barbican-keystone-listener-log/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.221666 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7fffd76679-n928f_ec0b4c68-f07d-4d42-8225-6950e9315970/barbican-worker-log/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.410173 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-bz9xk_938c7955-b34d-435c-8933-2c3dc8043578/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.465417 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/ceilometer-central-agent/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.596905 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/ceilometer-notification-agent/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.655861 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/proxy-httpd/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.657585 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41dfcd37-c0b9-426a-bbf9-fca697e2368a/sg-core/0.log" Oct 14 10:06:33 crc kubenswrapper[5002]: I1014 10:06:33.842215 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-68t9r_d93bac7d-9b37-4cfe-98dd-271aeff235ff/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.008072 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-vpb2h_b871be8d-75a8-4c47-bfe4-4ad8f79d7fec/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.166026 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d29ae9ac-13f1-4621-9696-4eabea39109f/cinder-api-log/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.250984 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d29ae9ac-13f1-4621-9696-4eabea39109f/cinder-api/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.487110 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_947933b0-5baf-4203-bea2-3b229c31ab91/probe/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.605247 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_947933b0-5baf-4203-bea2-3b229c31ab91/cinder-backup/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.706487 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3f1dbab-8a06-4b99-8f41-39179983cd6c/cinder-scheduler/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.791055 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f3f1dbab-8a06-4b99-8f41-39179983cd6c/probe/0.log" Oct 14 10:06:34 crc kubenswrapper[5002]: I1014 10:06:34.959748 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a/cinder-volume/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.041394 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a58e0c7f-ecf1-4de2-9298-eb1f9ed6743a/probe/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.126461 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-g5mq8_a6cb9d07-96cd-42d8-91a1-230e4f6316f0/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.256702 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-zhs2d_34674ceb-4ae7-48fb-84be-afaae15ca5bf/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.499057 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/init/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.629123 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/init/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.728629 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0b995566-e3f1-4c1c-a9e0-02874deed49b/glance-httpd/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.829061 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0b995566-e3f1-4c1c-a9e0-02874deed49b/glance-log/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.863346 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d78c4b9-jm4l9_244cc8e3-294e-4c7f-b77f-10596f9e56ab/dnsmasq-dns/0.log" Oct 14 10:06:35 crc kubenswrapper[5002]: I1014 10:06:35.991701 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f64389b8-a21d-4cf5-abe8-402b170db4b3/glance-httpd/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.033117 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f64389b8-a21d-4cf5-abe8-402b170db4b3/glance-log/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.284176 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-557d55dc9d-sk8l2_681029aa-9883-4641-9008-aeb673358cdd/horizon/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.300804 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_df77ae37-bb4c-433c-bd13-bc48e1e6fa2c/horizontest-tests-horizontest/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.544086 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-6jrqf_1bbe90c4-4bf2-4f7e-ae65-8ac5a5324819/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.794772 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sfkxq_f8e84847-a7b7-43a3-90c7-7a9c1ea92415/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:36 crc kubenswrapper[5002]: I1014 10:06:36.993021 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29340541-twn6l_181a00bb-d0d0-483b-a51e-a84e308f06c8/keystone-cron/0.log" Oct 14 10:06:37 crc kubenswrapper[5002]: I1014 10:06:37.202923 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29340601-5wkbp_4d2fdadc-5e2c-4186-8402-229511267bb2/keystone-cron/0.log" Oct 14 10:06:37 crc kubenswrapper[5002]: I1014 10:06:37.471184 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_670693df-bbd2-4bf4-a705-8cc8a02d3d2f/kube-state-metrics/0.log" Oct 14 10:06:37 crc kubenswrapper[5002]: I1014 10:06:37.644412 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-557d55dc9d-sk8l2_681029aa-9883-4641-9008-aeb673358cdd/horizon-log/0.log" Oct 14 10:06:37 crc kubenswrapper[5002]: I1014 10:06:37.679953 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-cwbrj_4d34c041-ecf1-4380-8fd9-6bd9a9660f87/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:37 crc kubenswrapper[5002]: I1014 10:06:37.917819 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_978c2b52-ead1-4ca3-924e-83ed4bb321a5/manila-api-log/0.log" Oct 14 10:06:38 crc kubenswrapper[5002]: I1014 10:06:38.116218 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_978c2b52-ead1-4ca3-924e-83ed4bb321a5/manila-api/0.log" Oct 14 10:06:38 crc kubenswrapper[5002]: I1014 10:06:38.232412 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_0bdd233c-7fae-4db7-8056-a218f6da8754/manila-scheduler/0.log" Oct 14 10:06:38 crc kubenswrapper[5002]: I1014 10:06:38.317701 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_0bdd233c-7fae-4db7-8056-a218f6da8754/probe/0.log" Oct 14 10:06:38 crc kubenswrapper[5002]: I1014 10:06:38.542263 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d10fce14-aa5a-466f-948a-7787f2aa8a76/probe/0.log" Oct 14 10:06:38 crc kubenswrapper[5002]: I1014 10:06:38.551212 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_d10fce14-aa5a-466f-948a-7787f2aa8a76/manila-share/0.log" Oct 14 10:06:39 crc kubenswrapper[5002]: I1014 10:06:39.380143 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7db98f685f-fphzv_e587d08e-64fa-4e35-98dc-ee283ada8c7f/keystone-api/0.log" Oct 14 10:06:39 crc kubenswrapper[5002]: I1014 10:06:39.779023 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-d9t7b_721fe1de-3687-4a8c-babb-e30f9ec46abe/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:39 crc kubenswrapper[5002]: I1014 10:06:39.871647 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7855f7b865-dtwlh_abe829e6-d42f-447f-bcb2-aad97c1e4f92/neutron-httpd/0.log" Oct 14 10:06:40 crc kubenswrapper[5002]: I1014 10:06:40.355690 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7855f7b865-dtwlh_abe829e6-d42f-447f-bcb2-aad97c1e4f92/neutron-api/0.log" Oct 14 10:06:41 crc kubenswrapper[5002]: I1014 10:06:41.474589 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fc8e7450-29ac-4acb-a090-b7b0d1f2666f/nova-cell0-conductor-conductor/0.log" Oct 14 10:06:42 crc kubenswrapper[5002]: I1014 10:06:42.162230 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ba1b8fb3-2b74-47f6-adfe-58a6084311fb/nova-cell1-conductor-conductor/0.log" Oct 14 10:06:42 crc kubenswrapper[5002]: I1014 10:06:42.919364 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9b0062d-6204-46b8-8614-26f00ff2efd7/nova-cell1-novncproxy-novncproxy/0.log" Oct 14 10:06:43 crc kubenswrapper[5002]: I1014 10:06:43.443184 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8tsg5_99e9a85b-e7b7-4f1d-89e8-ea6b5c86efb2/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:43 crc kubenswrapper[5002]: I1014 10:06:43.723189 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55d153d2-2f4c-4e77-9a7a-fa120ac94da2/nova-api-log/0.log" Oct 14 10:06:43 crc kubenswrapper[5002]: I1014 10:06:43.897380 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5/nova-metadata-log/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.005645 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_59facc52-a7a5-4a65-b7ba-67e9d8267de2/nova-scheduler-scheduler/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.276860 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55d153d2-2f4c-4e77-9a7a-fa120ac94da2/nova-api-api/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.514166 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/mysql-bootstrap/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.643004 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/mysql-bootstrap/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.686445 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2c1f1d79-cecd-4eda-a876-11dfb83ab76b/galera/0.log" Oct 14 10:06:45 crc kubenswrapper[5002]: I1014 10:06:45.880791 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/mysql-bootstrap/0.log" Oct 14 10:06:46 crc kubenswrapper[5002]: I1014 10:06:46.165196 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/mysql-bootstrap/0.log" Oct 14 10:06:46 crc kubenswrapper[5002]: I1014 10:06:46.218859 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_355bb6ac-9bd2-411e-81d4-400069911641/galera/0.log" Oct 14 10:06:46 crc kubenswrapper[5002]: I1014 10:06:46.418259 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c0df2486-02ae-48b5-b3b9-acb14ec03a6f/openstackclient/0.log" Oct 14 10:06:46 crc kubenswrapper[5002]: I1014 10:06:46.602585 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-b85t2_8622c6ea-480a-457b-b42c-2adad7acc425/openstack-network-exporter/0.log" Oct 14 10:06:46 crc kubenswrapper[5002]: I1014 10:06:46.870035 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server-init/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.023324 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server-init/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.090537 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovs-vswitchd/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.208472 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-n5j7z_4074edb0-c770-4819-a4bd-581f3e7c6e23/ovsdb-server/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.427469 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tw5d7_f9439452-e42f-4fa8-ac61-65b5389ff828/ovn-controller/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.640653 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e39c43cd-f4d2-40fd-82b5-fa2d570ac1d5/nova-metadata-metadata/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.660271 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-sz55g_0b20c844-7dba-4cdf-b6a2-b7deff3e401a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.835299 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3b8bf477-7229-46cd-aabc-2ae7794a694b/openstack-network-exporter/0.log" Oct 14 10:06:47 crc kubenswrapper[5002]: I1014 10:06:47.855277 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3b8bf477-7229-46cd-aabc-2ae7794a694b/ovn-northd/0.log" Oct 14 10:06:48 crc kubenswrapper[5002]: I1014 10:06:48.057138 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_705e9cf2-0759-49ed-8407-a73c3aed75cf/openstack-network-exporter/0.log" Oct 14 10:06:48 crc kubenswrapper[5002]: I1014 10:06:48.099266 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_705e9cf2-0759-49ed-8407-a73c3aed75cf/ovsdbserver-nb/0.log" Oct 14 10:06:48 crc kubenswrapper[5002]: I1014 10:06:48.290244 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2b959df9-bb11-4ee8-8af1-73beac51cfba/openstack-network-exporter/0.log" Oct 14 10:06:48 crc kubenswrapper[5002]: I1014 10:06:48.316165 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_2b959df9-bb11-4ee8-8af1-73beac51cfba/ovsdbserver-sb/0.log" Oct 14 10:06:48 crc kubenswrapper[5002]: I1014 10:06:48.895913 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/setup-container/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.016879 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5dc947dd58-kwj5n_0141dd41-3526-47f8-8477-f0d347c58cea/placement-api/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.134733 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/setup-container/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.170782 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5dc947dd58-kwj5n_0141dd41-3526-47f8-8477-f0d347c58cea/placement-log/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.217316 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_94673e4e-a0de-4a23-981d-1a6c7ea055ff/rabbitmq/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.383850 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/setup-container/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.559134 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/setup-container/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.577215 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_bcf9ff36-3590-4122-881c-6331338dd63f/rabbitmq/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.754138 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-f955z_3b28c2e3-928a-400a-92e3-c4e407257f94/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:49 crc kubenswrapper[5002]: I1014 10:06:49.846449 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vmqs2_fc36bcdf-48eb-4ef8-90c9-eb7842529652/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.071151 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-hlm2l_75c3bce2-2453-4d20-a946-9a28fe2151e8/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.199116 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-46bxg_b1704867-21c5-46a1-8d37-85af1337b303/ssh-known-hosts-edpm-deployment/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.450462 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_ec2597cf-678d-44a5-a3e3-41be976a3dc6/tempest-tests-tempest-tests-runner/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.522579 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_08748a2d-35ba-47e7-a01a-51f71b61d041/tempest-tests-tempest-tests-runner/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.683362 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_487e1aee-dbf2-451c-86dd-e2df3f229d89/test-operator-logs-container/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.864413 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_845f0c8f-dc7f-4de3-8b0d-6fd3428c15f1/test-operator-logs-container/0.log" Oct 14 10:06:50 crc kubenswrapper[5002]: I1014 10:06:50.937981 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c8166676-1bb1-4eec-9d75-28ccb7357ec3/test-operator-logs-container/0.log" Oct 14 10:06:51 crc kubenswrapper[5002]: I1014 10:06:51.103863 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_96b65a05-43d2-4d66-b467-3c88b9ce57e4/test-operator-logs-container/0.log" Oct 14 10:06:51 crc kubenswrapper[5002]: I1014 10:06:51.334522 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_8c2247d7-13c4-4c8d-ab13-3f2fbd020a1e/tobiko-tests-tobiko/0.log" Oct 14 10:06:51 crc kubenswrapper[5002]: I1014 10:06:51.407028 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_d8f34f80-d17b-4cdb-9834-f2e96de161ae/tobiko-tests-tobiko/0.log" Oct 14 10:06:51 crc kubenswrapper[5002]: I1014 10:06:51.599863 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-fcrcn_7229bab8-aa3f-46dc-bcfc-338bce4801af/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 14 10:06:52 crc kubenswrapper[5002]: I1014 10:06:52.019270 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d0a98c43-f790-476c-822c-e9b5957bbb78/memcached/0.log" Oct 14 10:07:09 crc kubenswrapper[5002]: I1014 10:07:09.218753 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:07:09 crc kubenswrapper[5002]: I1014 10:07:09.219442 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:07:13 crc kubenswrapper[5002]: I1014 10:07:13.960727 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.110336 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.134992 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.151011 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.325363 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/util/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.374194 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/pull/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.374283 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036atzb5s_60f12264-cedf-4b85-8aa3-2707d34e6a00/extract/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.490514 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-qpnq7_46630983-31be-46df-9e05-0506abbaf29e/kube-rbac-proxy/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.577703 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lr6gb_cfe90f21-7b6a-4364-8c33-92b2880e9e0b/kube-rbac-proxy/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.604916 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-qpnq7_46630983-31be-46df-9e05-0506abbaf29e/manager/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.741111 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lr6gb_cfe90f21-7b6a-4364-8c33-92b2880e9e0b/manager/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.767811 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-9kfv7_578becbc-f063-4b98-b5a3-eb0656b24850/kube-rbac-proxy/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.797645 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-9kfv7_578becbc-f063-4b98-b5a3-eb0656b24850/manager/0.log" Oct 14 10:07:14 crc kubenswrapper[5002]: I1014 10:07:14.961309 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-8j2qn_de07a029-a377-47c8-be2f-a31148f4948a/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.021752 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-8j2qn_de07a029-a377-47c8-be2f-a31148f4948a/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.141734 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-5t4hm_63c73302-dc00-4dfe-9d3f-d80cb181ffc2/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.167880 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-5t4hm_63c73302-dc00-4dfe-9d3f-d80cb181ffc2/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.230916 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-s6vkc_bbc859e5-324a-413c-a7ac-499a0b0b8fba/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.320012 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-s6vkc_bbc859e5-324a-413c-a7ac-499a0b0b8fba/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.387172 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-d9ftz_17c37ba2-ff55-4b60-a1db-01dec0522e2e/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.555657 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-7jcbw_e1488db8-3d22-4c1f-8400-8d292b48027a/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.571427 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-d9ftz_17c37ba2-ff55-4b60-a1db-01dec0522e2e/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.588574 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-7jcbw_e1488db8-3d22-4c1f-8400-8d292b48027a/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.730781 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-kqdfb_4b405ac3-bbcd-4432-8f99-2862ef406541/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.792022 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-kqdfb_4b405ac3-bbcd-4432-8f99-2862ef406541/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.883530 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-59jzf_2041ec73-97de-43ea-ae4e-e8a1d7521157/kube-rbac-proxy/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.935094 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-59jzf_2041ec73-97de-43ea-ae4e-e8a1d7521157/manager/0.log" Oct 14 10:07:15 crc kubenswrapper[5002]: I1014 10:07:15.991443 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-xsb2d_57a0f853-8feb-4064-9ed2-3460f2f66901/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.119741 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-xsb2d_57a0f853-8feb-4064-9ed2-3460f2f66901/manager/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.181669 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-69rk7_95e654ed-72c1-4f5b-b670-a89c96692455/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.266935 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-69rk7_95e654ed-72c1-4f5b-b670-a89c96692455/manager/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.313181 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-82c6h_9c72f60f-c311-45eb-84a1-8c95164be376/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.450294 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-82c6h_9c72f60f-c311-45eb-84a1-8c95164be376/manager/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.516556 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-cn7pq_27e01872-38da-4f89-8232-c8edaec917f8/manager/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.547171 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-cn7pq_27e01872-38da-4f89-8232-c8edaec917f8/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.679446 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d44848n4xg8_05103074-be63-4240-bd7e-737c3c6ff4d5/manager/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.686070 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d44848n4xg8_05103074-be63-4240-bd7e-737c3c6ff4d5/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.849011 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-k6scp_516be58e-c541-4fbe-b4b8-127e07a0b8b7/kube-rbac-proxy/0.log" Oct 14 10:07:16 crc kubenswrapper[5002]: I1014 10:07:16.992683 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd554b564-6djql_41ec772c-a692-4e45-90f1-5466203d5587/kube-rbac-proxy/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.187574 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd554b564-6djql_41ec772c-a692-4e45-90f1-5466203d5587/operator/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.315011 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-pgk79_4f65b042-e91d-4c89-9aa4-52abc662f94f/registry-server/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.418919 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-xj9vc_8e0847dd-54d1-44f7-8df9-b1378b00cb67/kube-rbac-proxy/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.485549 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-xj9vc_8e0847dd-54d1-44f7-8df9-b1378b00cb67/manager/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.603616 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-tws79_20fefb60-1453-4a27-9f7a-518df5b902fc/kube-rbac-proxy/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.620498 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-tws79_20fefb60-1453-4a27-9f7a-518df5b902fc/manager/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.778052 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-s89dt_b1921718-5347-4bfd-a327-9c38671b4888/operator/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.805637 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-74c8t_b04b0d3d-03a0-4f12-814f-599f2899962b/kube-rbac-proxy/0.log" Oct 14 10:07:17 crc kubenswrapper[5002]: I1014 10:07:17.911705 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-74c8t_b04b0d3d-03a0-4f12-814f-599f2899962b/manager/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.040012 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-k6scp_516be58e-c541-4fbe-b4b8-127e07a0b8b7/manager/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.043105 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-fp4rw_42f32565-187e-4c71-a2c3-35eb988b8243/kube-rbac-proxy/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.165701 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-fp4rw_42f32565-187e-4c71-a2c3-35eb988b8243/manager/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.190122 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c475b9f8d-pbb6v_cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f/kube-rbac-proxy/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.277243 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-c475b9f8d-pbb6v_cdd8f7a8-ef6d-4166-b7d8-2bc5cfef0b7f/manager/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.331742 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-lwlwj_74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8/kube-rbac-proxy/0.log" Oct 14 10:07:18 crc kubenswrapper[5002]: I1014 10:07:18.386699 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-lwlwj_74ae3a64-3a2b-4cd8-bf3f-b724622ee1f8/manager/0.log" Oct 14 10:07:22 crc kubenswrapper[5002]: E1014 10:07:22.720344 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:07:34 crc kubenswrapper[5002]: I1014 10:07:34.774985 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9gbqd_6b840956-6900-4c7c-89ce-622ddf455cf6/control-plane-machine-set-operator/0.log" Oct 14 10:07:34 crc kubenswrapper[5002]: I1014 10:07:34.952335 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hjt8h_71022b1b-51c8-4c8c-8d57-dc220ebf1b05/kube-rbac-proxy/0.log" Oct 14 10:07:35 crc kubenswrapper[5002]: I1014 10:07:35.001485 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hjt8h_71022b1b-51c8-4c8c-8d57-dc220ebf1b05/machine-api-operator/0.log" Oct 14 10:07:39 crc kubenswrapper[5002]: I1014 10:07:39.218293 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:07:39 crc kubenswrapper[5002]: I1014 10:07:39.218920 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:07:48 crc kubenswrapper[5002]: I1014 10:07:48.033944 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hsnk9_cbaa9318-0355-41df-b7bd-bd2fe2ce7d30/cert-manager-controller/0.log" Oct 14 10:07:48 crc kubenswrapper[5002]: I1014 10:07:48.141660 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-wkw2k_632fb195-4a92-40b0-96e7-4db2227146d5/cert-manager-cainjector/0.log" Oct 14 10:07:48 crc kubenswrapper[5002]: I1014 10:07:48.219038 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-66k4z_067cf99a-23c5-4e71-85f5-1d2fbfb4ccb7/cert-manager-webhook/0.log" Oct 14 10:08:01 crc kubenswrapper[5002]: I1014 10:08:01.512296 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-s9gvm_59bfeb19-1a15-4cbd-8970-5c18ff149992/nmstate-console-plugin/0.log" Oct 14 10:08:01 crc kubenswrapper[5002]: I1014 10:08:01.676710 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-82rj9_0629a9d2-3340-4e00-9673-baaaf354a8ce/nmstate-handler/0.log" Oct 14 10:08:01 crc kubenswrapper[5002]: I1014 10:08:01.794063 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dq7k8_89f2e467-845f-45fb-b411-48c0cf2d3f0d/nmstate-metrics/0.log" Oct 14 10:08:01 crc kubenswrapper[5002]: I1014 10:08:01.794561 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dq7k8_89f2e467-845f-45fb-b411-48c0cf2d3f0d/kube-rbac-proxy/0.log" Oct 14 10:08:01 crc kubenswrapper[5002]: I1014 10:08:01.963053 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hpwzt_38ea0de3-d753-486f-8fc0-b5239e804cb2/nmstate-operator/0.log" Oct 14 10:08:02 crc kubenswrapper[5002]: I1014 10:08:02.004871 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-4tmr2_cb6927e5-34dd-43d1-9dea-db00339df27e/nmstate-webhook/0.log" Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.218260 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.218852 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.218906 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.219660 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.219712 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb" gracePeriod=600 Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.439328 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb" exitCode=0 Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.439376 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb"} Oct 14 10:08:09 crc kubenswrapper[5002]: I1014 10:08:09.439414 5002 scope.go:117] "RemoveContainer" containerID="499c8d4b482b155691ef2060587ee1875e0efbeea9acac729258559089a3173b" Oct 14 10:08:10 crc kubenswrapper[5002]: I1014 10:08:10.456586 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerStarted","Data":"66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7"} Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.529490 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2zwcb_0a00fbf0-fcf8-42f8-a8da-cbf266655ee3/kube-rbac-proxy/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.671742 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2zwcb_0a00fbf0-fcf8-42f8-a8da-cbf266655ee3/controller/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.757945 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.932869 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.955009 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.974205 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:08:16 crc kubenswrapper[5002]: I1014 10:08:16.976247 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.147420 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.176493 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.176936 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.257992 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.361224 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-reloader/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.381344 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-metrics/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.389642 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/cp-frr-files/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.439707 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/controller/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.572034 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/kube-rbac-proxy/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.612357 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/frr-metrics/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.638334 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/kube-rbac-proxy-frr/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.769370 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/reloader/0.log" Oct 14 10:08:17 crc kubenswrapper[5002]: I1014 10:08:17.793355 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-lf7b4_10d5d438-0656-4699-b386-a539ca6810d2/frr-k8s-webhook-server/0.log" Oct 14 10:08:18 crc kubenswrapper[5002]: I1014 10:08:18.090243 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7769bf444d-r6x5z_45cb6486-ddee-4491-86d3-6de31fa2eeed/manager/0.log" Oct 14 10:08:18 crc kubenswrapper[5002]: I1014 10:08:18.257136 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74b78f565b-6dh62_513044aa-d565-47b2-8fb7-5df37788a795/webhook-server/0.log" Oct 14 10:08:18 crc kubenswrapper[5002]: I1014 10:08:18.357007 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-686b6_6d62cfe7-6906-4f1f-bc0b-595c98d4af86/kube-rbac-proxy/0.log" Oct 14 10:08:19 crc kubenswrapper[5002]: I1014 10:08:19.048044 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-686b6_6d62cfe7-6906-4f1f-bc0b-595c98d4af86/speaker/0.log" Oct 14 10:08:19 crc kubenswrapper[5002]: I1014 10:08:19.586441 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-t5nqw_a31644e2-8ecc-4402-8776-2eb5815c7d55/frr/0.log" Oct 14 10:08:31 crc kubenswrapper[5002]: I1014 10:08:31.985281 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.142516 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.174217 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.206214 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.291317 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/pull/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.309364 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/util/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.338593 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d27pbzv_9a92c18c-a9d6-470f-9236-feab1741aa53/extract/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.446900 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.672851 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.716959 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.724025 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.849020 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-utilities/0.log" Oct 14 10:08:32 crc kubenswrapper[5002]: I1014 10:08:32.867648 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/extract-content/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.042157 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.222217 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.295580 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.348139 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.504132 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-content/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.504692 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/extract-utilities/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.763600 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.900393 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dzd5b_1cb8dd4f-12fd-46d1-99d3-61b66d6147a1/registry-server/0.log" Oct 14 10:08:33 crc kubenswrapper[5002]: I1014 10:08:33.988976 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.089941 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.211154 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.332929 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/util/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.355488 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/pull/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.405346 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cxvsdv_909ae76b-abca-406b-9389-49b934d5f1fc/extract/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.586085 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-cq8mp_651e3f67-0cad-4654-8c98-0672dd56f367/marketplace-operator/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.714323 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v6wmt_1d5a10a8-19fd-46f5-8062-d3e1db89f218/registry-server/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.817509 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.961857 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:08:34 crc kubenswrapper[5002]: I1014 10:08:34.987828 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.015436 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.144417 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-utilities/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.194962 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/extract-content/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.350853 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.359964 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nggtm_7f04708c-e386-4cfa-979f-9b930ee896ae/registry-server/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.591439 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.597053 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.606939 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.799716 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-content/0.log" Oct 14 10:08:35 crc kubenswrapper[5002]: I1014 10:08:35.818737 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/extract-utilities/0.log" Oct 14 10:08:36 crc kubenswrapper[5002]: I1014 10:08:36.741405 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ts5dh_ae36f67a-ccd9-47f8-8030-6e1eb0fbcb15/registry-server/0.log" Oct 14 10:08:45 crc kubenswrapper[5002]: I1014 10:08:45.937053 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:45 crc kubenswrapper[5002]: E1014 10:08:45.938216 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" containerName="container-00" Oct 14 10:08:45 crc kubenswrapper[5002]: I1014 10:08:45.938234 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" containerName="container-00" Oct 14 10:08:45 crc kubenswrapper[5002]: I1014 10:08:45.938486 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc47e8d-d884-4ecc-a1ee-c0acf7c12f00" containerName="container-00" Oct 14 10:08:45 crc kubenswrapper[5002]: I1014 10:08:45.940270 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:45 crc kubenswrapper[5002]: I1014 10:08:45.955597 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.056874 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.057176 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.057227 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xnh\" (UniqueName: \"kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.158894 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.158990 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.159079 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xnh\" (UniqueName: \"kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.160266 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.160680 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.185888 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xnh\" (UniqueName: \"kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh\") pod \"certified-operators-fv5k8\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.276388 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:46 crc kubenswrapper[5002]: I1014 10:08:46.820508 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:46 crc kubenswrapper[5002]: W1014 10:08:46.827009 5002 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod773c4ee9_d141_47da_8676_1f9f43e44ce4.slice/crio-5c0638404d1afcf5bb07b685ff1e4161ac8403fbb95f93c611373a0c0cece20f WatchSource:0}: Error finding container 5c0638404d1afcf5bb07b685ff1e4161ac8403fbb95f93c611373a0c0cece20f: Status 404 returned error can't find the container with id 5c0638404d1afcf5bb07b685ff1e4161ac8403fbb95f93c611373a0c0cece20f Oct 14 10:08:47 crc kubenswrapper[5002]: I1014 10:08:47.797002 5002 generic.go:334] "Generic (PLEG): container finished" podID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerID="de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51" exitCode=0 Oct 14 10:08:47 crc kubenswrapper[5002]: I1014 10:08:47.797073 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerDied","Data":"de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51"} Oct 14 10:08:47 crc kubenswrapper[5002]: I1014 10:08:47.797363 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerStarted","Data":"5c0638404d1afcf5bb07b685ff1e4161ac8403fbb95f93c611373a0c0cece20f"} Oct 14 10:08:47 crc kubenswrapper[5002]: I1014 10:08:47.800804 5002 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:08:48 crc kubenswrapper[5002]: E1014 10:08:48.721071 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:08:48 crc kubenswrapper[5002]: I1014 10:08:48.806750 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerStarted","Data":"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2"} Oct 14 10:08:49 crc kubenswrapper[5002]: I1014 10:08:49.818912 5002 generic.go:334] "Generic (PLEG): container finished" podID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerID="6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2" exitCode=0 Oct 14 10:08:49 crc kubenswrapper[5002]: I1014 10:08:49.818958 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerDied","Data":"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2"} Oct 14 10:08:50 crc kubenswrapper[5002]: I1014 10:08:50.830501 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerStarted","Data":"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed"} Oct 14 10:08:50 crc kubenswrapper[5002]: I1014 10:08:50.851284 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fv5k8" podStartSLOduration=3.241601687 podStartE2EDuration="5.851266215s" podCreationTimestamp="2025-10-14 10:08:45 +0000 UTC" firstStartedPulling="2025-10-14 10:08:47.80052666 +0000 UTC m=+8260.781766122" lastFinishedPulling="2025-10-14 10:08:50.410191188 +0000 UTC m=+8263.391430650" observedRunningTime="2025-10-14 10:08:50.84770678 +0000 UTC m=+8263.828946242" watchObservedRunningTime="2025-10-14 10:08:50.851266215 +0000 UTC m=+8263.832505677" Oct 14 10:08:56 crc kubenswrapper[5002]: I1014 10:08:56.277456 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:56 crc kubenswrapper[5002]: I1014 10:08:56.278033 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:56 crc kubenswrapper[5002]: I1014 10:08:56.318730 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:56 crc kubenswrapper[5002]: I1014 10:08:56.934458 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:56 crc kubenswrapper[5002]: I1014 10:08:56.997810 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:58 crc kubenswrapper[5002]: I1014 10:08:58.907566 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fv5k8" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="registry-server" containerID="cri-o://44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed" gracePeriod=2 Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.419995 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.520074 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2xnh\" (UniqueName: \"kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh\") pod \"773c4ee9-d141-47da-8676-1f9f43e44ce4\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.520323 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities\") pod \"773c4ee9-d141-47da-8676-1f9f43e44ce4\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.520367 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content\") pod \"773c4ee9-d141-47da-8676-1f9f43e44ce4\" (UID: \"773c4ee9-d141-47da-8676-1f9f43e44ce4\") " Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.520895 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities" (OuterVolumeSpecName: "utilities") pod "773c4ee9-d141-47da-8676-1f9f43e44ce4" (UID: "773c4ee9-d141-47da-8676-1f9f43e44ce4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.524908 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh" (OuterVolumeSpecName: "kube-api-access-b2xnh") pod "773c4ee9-d141-47da-8676-1f9f43e44ce4" (UID: "773c4ee9-d141-47da-8676-1f9f43e44ce4"). InnerVolumeSpecName "kube-api-access-b2xnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.576085 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "773c4ee9-d141-47da-8676-1f9f43e44ce4" (UID: "773c4ee9-d141-47da-8676-1f9f43e44ce4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.622568 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2xnh\" (UniqueName: \"kubernetes.io/projected/773c4ee9-d141-47da-8676-1f9f43e44ce4-kube-api-access-b2xnh\") on node \"crc\" DevicePath \"\"" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.622601 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.622611 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/773c4ee9-d141-47da-8676-1f9f43e44ce4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.917191 5002 generic.go:334] "Generic (PLEG): container finished" podID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerID="44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed" exitCode=0 Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.917229 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerDied","Data":"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed"} Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.917256 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv5k8" event={"ID":"773c4ee9-d141-47da-8676-1f9f43e44ce4","Type":"ContainerDied","Data":"5c0638404d1afcf5bb07b685ff1e4161ac8403fbb95f93c611373a0c0cece20f"} Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.917271 5002 scope.go:117] "RemoveContainer" containerID="44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.917401 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv5k8" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.944222 5002 scope.go:117] "RemoveContainer" containerID="6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2" Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.944789 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.965141 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fv5k8"] Oct 14 10:08:59 crc kubenswrapper[5002]: I1014 10:08:59.977288 5002 scope.go:117] "RemoveContainer" containerID="de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.037587 5002 scope.go:117] "RemoveContainer" containerID="44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed" Oct 14 10:09:00 crc kubenswrapper[5002]: E1014 10:09:00.038388 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed\": container with ID starting with 44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed not found: ID does not exist" containerID="44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.038433 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed"} err="failed to get container status \"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed\": rpc error: code = NotFound desc = could not find container \"44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed\": container with ID starting with 44364bda6730829d0faece7d8ebe4cec2f54c95e54235466fd1619358f7689ed not found: ID does not exist" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.038458 5002 scope.go:117] "RemoveContainer" containerID="6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2" Oct 14 10:09:00 crc kubenswrapper[5002]: E1014 10:09:00.038817 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2\": container with ID starting with 6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2 not found: ID does not exist" containerID="6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.038860 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2"} err="failed to get container status \"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2\": rpc error: code = NotFound desc = could not find container \"6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2\": container with ID starting with 6a1e62549a8cc72c8d0e21002f28d1793760429261e400117638d9a3204608e2 not found: ID does not exist" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.038875 5002 scope.go:117] "RemoveContainer" containerID="de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51" Oct 14 10:09:00 crc kubenswrapper[5002]: E1014 10:09:00.039118 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51\": container with ID starting with de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51 not found: ID does not exist" containerID="de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51" Oct 14 10:09:00 crc kubenswrapper[5002]: I1014 10:09:00.039164 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51"} err="failed to get container status \"de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51\": rpc error: code = NotFound desc = could not find container \"de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51\": container with ID starting with de5eca0cfec069548f31ac382ada2f47e88c5ca4a66377ac5554dda57f09cf51 not found: ID does not exist" Oct 14 10:09:00 crc kubenswrapper[5002]: E1014 10:09:00.061597 5002 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.74:54526->38.102.83.74:44515: write tcp 38.102.83.74:54526->38.102.83.74:44515: write: broken pipe Oct 14 10:09:01 crc kubenswrapper[5002]: I1014 10:09:01.735010 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" path="/var/lib/kubelet/pods/773c4ee9-d141-47da-8676-1f9f43e44ce4/volumes" Oct 14 10:09:49 crc kubenswrapper[5002]: E1014 10:09:49.724912 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:10:09 crc kubenswrapper[5002]: I1014 10:10:09.218258 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:10:09 crc kubenswrapper[5002]: I1014 10:10:09.219267 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:10:39 crc kubenswrapper[5002]: I1014 10:10:39.218572 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:10:39 crc kubenswrapper[5002]: I1014 10:10:39.219614 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:10:43 crc kubenswrapper[5002]: I1014 10:10:43.164181 5002 generic.go:334] "Generic (PLEG): container finished" podID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerID="718a01b93b7cb686b6d41dff767e8cd66e65b6de0efefc85f5c1a7267467f730" exitCode=0 Oct 14 10:10:43 crc kubenswrapper[5002]: I1014 10:10:43.164291 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-k2pt5/must-gather-2dtds" event={"ID":"3eb989ee-6138-47c7-8d09-dedcf1364fd2","Type":"ContainerDied","Data":"718a01b93b7cb686b6d41dff767e8cd66e65b6de0efefc85f5c1a7267467f730"} Oct 14 10:10:43 crc kubenswrapper[5002]: I1014 10:10:43.165660 5002 scope.go:117] "RemoveContainer" containerID="718a01b93b7cb686b6d41dff767e8cd66e65b6de0efefc85f5c1a7267467f730" Oct 14 10:10:43 crc kubenswrapper[5002]: I1014 10:10:43.363481 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k2pt5_must-gather-2dtds_3eb989ee-6138-47c7-8d09-dedcf1364fd2/gather/0.log" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.989363 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:10:51 crc kubenswrapper[5002]: E1014 10:10:51.990336 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="extract-content" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.990349 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="extract-content" Oct 14 10:10:51 crc kubenswrapper[5002]: E1014 10:10:51.990371 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="registry-server" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.990377 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="registry-server" Oct 14 10:10:51 crc kubenswrapper[5002]: E1014 10:10:51.990397 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="extract-utilities" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.990403 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="extract-utilities" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.990602 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="773c4ee9-d141-47da-8676-1f9f43e44ce4" containerName="registry-server" Oct 14 10:10:51 crc kubenswrapper[5002]: I1014 10:10:51.991875 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.008080 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.181245 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5gtk\" (UniqueName: \"kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.181700 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.181744 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.283848 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.283916 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.284051 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5gtk\" (UniqueName: \"kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.284451 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.284800 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.310302 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5gtk\" (UniqueName: \"kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk\") pod \"redhat-operators-lljdl\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.318609 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:10:52 crc kubenswrapper[5002]: I1014 10:10:52.843967 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:10:53 crc kubenswrapper[5002]: I1014 10:10:53.269372 5002 generic.go:334] "Generic (PLEG): container finished" podID="8866cded-95f4-4436-add4-5465ad800328" containerID="4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508" exitCode=0 Oct 14 10:10:53 crc kubenswrapper[5002]: I1014 10:10:53.269595 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerDied","Data":"4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508"} Oct 14 10:10:53 crc kubenswrapper[5002]: I1014 10:10:53.269641 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerStarted","Data":"0fe9bc2abe7a160ae9f0d34ff67edaad5a239a7e71ed0f3e5fc4518b9ae10cf0"} Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.393562 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.406511 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.416202 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.532425 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xspz6\" (UniqueName: \"kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.532714 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.532773 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.634947 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.634991 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.635116 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xspz6\" (UniqueName: \"kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.635577 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.635958 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.681357 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xspz6\" (UniqueName: \"kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6\") pod \"redhat-marketplace-kptqn\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:54 crc kubenswrapper[5002]: I1014 10:10:54.740987 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.193533 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.311099 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerStarted","Data":"f5b45b1c2fcead1d8f6775bef392a5a34bca130468fcc039dcd7842903512440"} Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.322061 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerStarted","Data":"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6"} Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.910002 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-k2pt5/must-gather-2dtds"] Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.910284 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-k2pt5/must-gather-2dtds" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="copy" containerID="cri-o://f346cc4a1dc60de31a83a0d527785353d09b17b196d072ffb455d987700d1f54" gracePeriod=2 Oct 14 10:10:55 crc kubenswrapper[5002]: I1014 10:10:55.922120 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-k2pt5/must-gather-2dtds"] Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.330343 5002 generic.go:334] "Generic (PLEG): container finished" podID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerID="d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724" exitCode=0 Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.330694 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerDied","Data":"d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724"} Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.336726 5002 generic.go:334] "Generic (PLEG): container finished" podID="8866cded-95f4-4436-add4-5465ad800328" containerID="744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6" exitCode=0 Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.336788 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerDied","Data":"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6"} Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.343289 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k2pt5_must-gather-2dtds_3eb989ee-6138-47c7-8d09-dedcf1364fd2/copy/0.log" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.344361 5002 generic.go:334] "Generic (PLEG): container finished" podID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerID="f346cc4a1dc60de31a83a0d527785353d09b17b196d072ffb455d987700d1f54" exitCode=143 Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.344413 5002 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc41322aa2c44c4d19f4ff290a85c567ac44f54bc682c95c516398327c60e441" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.385403 5002 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-k2pt5_must-gather-2dtds_3eb989ee-6138-47c7-8d09-dedcf1364fd2/copy/0.log" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.385682 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.481656 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output\") pod \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.481986 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxdxj\" (UniqueName: \"kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj\") pod \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\" (UID: \"3eb989ee-6138-47c7-8d09-dedcf1364fd2\") " Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.490156 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj" (OuterVolumeSpecName: "kube-api-access-zxdxj") pod "3eb989ee-6138-47c7-8d09-dedcf1364fd2" (UID: "3eb989ee-6138-47c7-8d09-dedcf1364fd2"). InnerVolumeSpecName "kube-api-access-zxdxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.585756 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxdxj\" (UniqueName: \"kubernetes.io/projected/3eb989ee-6138-47c7-8d09-dedcf1364fd2-kube-api-access-zxdxj\") on node \"crc\" DevicePath \"\"" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.653492 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3eb989ee-6138-47c7-8d09-dedcf1364fd2" (UID: "3eb989ee-6138-47c7-8d09-dedcf1364fd2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:10:56 crc kubenswrapper[5002]: I1014 10:10:56.690552 5002 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3eb989ee-6138-47c7-8d09-dedcf1364fd2-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 10:10:57 crc kubenswrapper[5002]: I1014 10:10:57.355226 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-k2pt5/must-gather-2dtds" Oct 14 10:10:57 crc kubenswrapper[5002]: I1014 10:10:57.733997 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" path="/var/lib/kubelet/pods/3eb989ee-6138-47c7-8d09-dedcf1364fd2/volumes" Oct 14 10:10:58 crc kubenswrapper[5002]: I1014 10:10:58.385070 5002 generic.go:334] "Generic (PLEG): container finished" podID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerID="d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d" exitCode=0 Oct 14 10:10:58 crc kubenswrapper[5002]: I1014 10:10:58.385510 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerDied","Data":"d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d"} Oct 14 10:10:58 crc kubenswrapper[5002]: I1014 10:10:58.398894 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerStarted","Data":"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2"} Oct 14 10:10:58 crc kubenswrapper[5002]: I1014 10:10:58.435818 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lljdl" podStartSLOduration=3.477844084 podStartE2EDuration="7.435796831s" podCreationTimestamp="2025-10-14 10:10:51 +0000 UTC" firstStartedPulling="2025-10-14 10:10:53.272244424 +0000 UTC m=+8386.253483876" lastFinishedPulling="2025-10-14 10:10:57.230197131 +0000 UTC m=+8390.211436623" observedRunningTime="2025-10-14 10:10:58.430569311 +0000 UTC m=+8391.411808763" watchObservedRunningTime="2025-10-14 10:10:58.435796831 +0000 UTC m=+8391.417036283" Oct 14 10:10:59 crc kubenswrapper[5002]: I1014 10:10:59.411403 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerStarted","Data":"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28"} Oct 14 10:10:59 crc kubenswrapper[5002]: I1014 10:10:59.458947 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kptqn" podStartSLOduration=3.025355441 podStartE2EDuration="5.458919717s" podCreationTimestamp="2025-10-14 10:10:54 +0000 UTC" firstStartedPulling="2025-10-14 10:10:56.334357882 +0000 UTC m=+8389.315597334" lastFinishedPulling="2025-10-14 10:10:58.767922148 +0000 UTC m=+8391.749161610" observedRunningTime="2025-10-14 10:10:59.436466838 +0000 UTC m=+8392.417706340" watchObservedRunningTime="2025-10-14 10:10:59.458919717 +0000 UTC m=+8392.440159199" Oct 14 10:11:02 crc kubenswrapper[5002]: I1014 10:11:02.319206 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:02 crc kubenswrapper[5002]: I1014 10:11:02.321593 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:03 crc kubenswrapper[5002]: I1014 10:11:03.380526 5002 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lljdl" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="registry-server" probeResult="failure" output=< Oct 14 10:11:03 crc kubenswrapper[5002]: timeout: failed to connect service ":50051" within 1s Oct 14 10:11:03 crc kubenswrapper[5002]: > Oct 14 10:11:04 crc kubenswrapper[5002]: I1014 10:11:04.741212 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:04 crc kubenswrapper[5002]: I1014 10:11:04.741286 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:04 crc kubenswrapper[5002]: I1014 10:11:04.837098 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:05 crc kubenswrapper[5002]: I1014 10:11:05.584238 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:05 crc kubenswrapper[5002]: I1014 10:11:05.666248 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:11:07 crc kubenswrapper[5002]: I1014 10:11:07.524720 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kptqn" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="registry-server" containerID="cri-o://e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28" gracePeriod=2 Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.083549 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.200777 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities\") pod \"81e2c1ff-9501-434c-b3d6-97a709c93ede\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.201178 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xspz6\" (UniqueName: \"kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6\") pod \"81e2c1ff-9501-434c-b3d6-97a709c93ede\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.201271 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content\") pod \"81e2c1ff-9501-434c-b3d6-97a709c93ede\" (UID: \"81e2c1ff-9501-434c-b3d6-97a709c93ede\") " Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.202829 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities" (OuterVolumeSpecName: "utilities") pod "81e2c1ff-9501-434c-b3d6-97a709c93ede" (UID: "81e2c1ff-9501-434c-b3d6-97a709c93ede"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.211679 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6" (OuterVolumeSpecName: "kube-api-access-xspz6") pod "81e2c1ff-9501-434c-b3d6-97a709c93ede" (UID: "81e2c1ff-9501-434c-b3d6-97a709c93ede"). InnerVolumeSpecName "kube-api-access-xspz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.216129 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81e2c1ff-9501-434c-b3d6-97a709c93ede" (UID: "81e2c1ff-9501-434c-b3d6-97a709c93ede"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.303801 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xspz6\" (UniqueName: \"kubernetes.io/projected/81e2c1ff-9501-434c-b3d6-97a709c93ede-kube-api-access-xspz6\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.303850 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.303864 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e2c1ff-9501-434c-b3d6-97a709c93ede-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.534822 5002 generic.go:334] "Generic (PLEG): container finished" podID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerID="e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28" exitCode=0 Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.534887 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerDied","Data":"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28"} Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.534914 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kptqn" event={"ID":"81e2c1ff-9501-434c-b3d6-97a709c93ede","Type":"ContainerDied","Data":"f5b45b1c2fcead1d8f6775bef392a5a34bca130468fcc039dcd7842903512440"} Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.534922 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kptqn" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.534929 5002 scope.go:117] "RemoveContainer" containerID="e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.571188 5002 scope.go:117] "RemoveContainer" containerID="d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.573608 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.590109 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kptqn"] Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.610160 5002 scope.go:117] "RemoveContainer" containerID="d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.674557 5002 scope.go:117] "RemoveContainer" containerID="e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28" Oct 14 10:11:08 crc kubenswrapper[5002]: E1014 10:11:08.675080 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28\": container with ID starting with e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28 not found: ID does not exist" containerID="e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.675112 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28"} err="failed to get container status \"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28\": rpc error: code = NotFound desc = could not find container \"e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28\": container with ID starting with e6327a7e5c2a500a5c5788b9036cf3987227256675586de9c342fc9b21a3bb28 not found: ID does not exist" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.675131 5002 scope.go:117] "RemoveContainer" containerID="d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d" Oct 14 10:11:08 crc kubenswrapper[5002]: E1014 10:11:08.675619 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d\": container with ID starting with d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d not found: ID does not exist" containerID="d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.675667 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d"} err="failed to get container status \"d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d\": rpc error: code = NotFound desc = could not find container \"d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d\": container with ID starting with d34ab1e286af034db8183f5ed4e058277633ca9dad65fa0296cbcd686ebfc36d not found: ID does not exist" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.675696 5002 scope.go:117] "RemoveContainer" containerID="d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724" Oct 14 10:11:08 crc kubenswrapper[5002]: E1014 10:11:08.676087 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724\": container with ID starting with d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724 not found: ID does not exist" containerID="d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724" Oct 14 10:11:08 crc kubenswrapper[5002]: I1014 10:11:08.676115 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724"} err="failed to get container status \"d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724\": rpc error: code = NotFound desc = could not find container \"d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724\": container with ID starting with d54ae7e387cdceb01fcc98b9752ffc18fd51e85c8a139dd4922a969dd8229724 not found: ID does not exist" Oct 14 10:11:08 crc kubenswrapper[5002]: E1014 10:11:08.724473 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.218739 5002 patch_prober.go:28] interesting pod/machine-config-daemon-7btf2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.219344 5002 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.219528 5002 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.220669 5002 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7"} pod="openshift-machine-config-operator/machine-config-daemon-7btf2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.220928 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerName="machine-config-daemon" containerID="cri-o://66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" gracePeriod=600 Oct 14 10:11:09 crc kubenswrapper[5002]: E1014 10:11:09.359375 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.549486 5002 generic.go:334] "Generic (PLEG): container finished" podID="6f3de631-7681-4103-8510-2b5545bbe8cd" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" exitCode=0 Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.549561 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" event={"ID":"6f3de631-7681-4103-8510-2b5545bbe8cd","Type":"ContainerDied","Data":"66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7"} Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.549625 5002 scope.go:117] "RemoveContainer" containerID="7ece999904ca952750f70e20d225a52d6b0dd4455fa11b177021b52b2640f0bb" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.554688 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:11:09 crc kubenswrapper[5002]: E1014 10:11:09.555728 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:11:09 crc kubenswrapper[5002]: I1014 10:11:09.734106 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" path="/var/lib/kubelet/pods/81e2c1ff-9501-434c-b3d6-97a709c93ede/volumes" Oct 14 10:11:12 crc kubenswrapper[5002]: I1014 10:11:12.398541 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:12 crc kubenswrapper[5002]: I1014 10:11:12.488058 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:12 crc kubenswrapper[5002]: I1014 10:11:12.648086 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:11:13 crc kubenswrapper[5002]: I1014 10:11:13.609008 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lljdl" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="registry-server" containerID="cri-o://c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2" gracePeriod=2 Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.067103 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.235648 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities\") pod \"8866cded-95f4-4436-add4-5465ad800328\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.235896 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content\") pod \"8866cded-95f4-4436-add4-5465ad800328\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.235928 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5gtk\" (UniqueName: \"kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk\") pod \"8866cded-95f4-4436-add4-5465ad800328\" (UID: \"8866cded-95f4-4436-add4-5465ad800328\") " Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.238959 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities" (OuterVolumeSpecName: "utilities") pod "8866cded-95f4-4436-add4-5465ad800328" (UID: "8866cded-95f4-4436-add4-5465ad800328"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.246125 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk" (OuterVolumeSpecName: "kube-api-access-k5gtk") pod "8866cded-95f4-4436-add4-5465ad800328" (UID: "8866cded-95f4-4436-add4-5465ad800328"). InnerVolumeSpecName "kube-api-access-k5gtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.329770 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8866cded-95f4-4436-add4-5465ad800328" (UID: "8866cded-95f4-4436-add4-5465ad800328"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.338945 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.339109 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5gtk\" (UniqueName: \"kubernetes.io/projected/8866cded-95f4-4436-add4-5465ad800328-kube-api-access-k5gtk\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.339176 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8866cded-95f4-4436-add4-5465ad800328-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.623088 5002 generic.go:334] "Generic (PLEG): container finished" podID="8866cded-95f4-4436-add4-5465ad800328" containerID="c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2" exitCode=0 Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.624303 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerDied","Data":"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2"} Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.624479 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lljdl" event={"ID":"8866cded-95f4-4436-add4-5465ad800328","Type":"ContainerDied","Data":"0fe9bc2abe7a160ae9f0d34ff67edaad5a239a7e71ed0f3e5fc4518b9ae10cf0"} Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.624627 5002 scope.go:117] "RemoveContainer" containerID="c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.624965 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lljdl" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.656337 5002 scope.go:117] "RemoveContainer" containerID="744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.696761 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.715786 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lljdl"] Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.716083 5002 scope.go:117] "RemoveContainer" containerID="4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.735922 5002 scope.go:117] "RemoveContainer" containerID="c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2" Oct 14 10:11:14 crc kubenswrapper[5002]: E1014 10:11:14.747462 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2\": container with ID starting with c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2 not found: ID does not exist" containerID="c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.747522 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2"} err="failed to get container status \"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2\": rpc error: code = NotFound desc = could not find container \"c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2\": container with ID starting with c8966a2950061630fa2ffb853796ca55d6252c64b94a7d12933b2fded79b83c2 not found: ID does not exist" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.747552 5002 scope.go:117] "RemoveContainer" containerID="744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6" Oct 14 10:11:14 crc kubenswrapper[5002]: E1014 10:11:14.748037 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6\": container with ID starting with 744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6 not found: ID does not exist" containerID="744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.749039 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6"} err="failed to get container status \"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6\": rpc error: code = NotFound desc = could not find container \"744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6\": container with ID starting with 744775da5e8ad0a43915780c9a48ff55270fad62069f0f0ee1d28ffb87f62ce6 not found: ID does not exist" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.749096 5002 scope.go:117] "RemoveContainer" containerID="4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508" Oct 14 10:11:14 crc kubenswrapper[5002]: E1014 10:11:14.749747 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508\": container with ID starting with 4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508 not found: ID does not exist" containerID="4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508" Oct 14 10:11:14 crc kubenswrapper[5002]: I1014 10:11:14.749788 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508"} err="failed to get container status \"4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508\": rpc error: code = NotFound desc = could not find container \"4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508\": container with ID starting with 4e3ebce1725ac87bd660feabde7b89a13280e4f9261b537fb0da480f6e060508 not found: ID does not exist" Oct 14 10:11:15 crc kubenswrapper[5002]: I1014 10:11:15.741577 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8866cded-95f4-4436-add4-5465ad800328" path="/var/lib/kubelet/pods/8866cded-95f4-4436-add4-5465ad800328/volumes" Oct 14 10:11:21 crc kubenswrapper[5002]: I1014 10:11:21.721139 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:11:21 crc kubenswrapper[5002]: E1014 10:11:21.722477 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:11:31 crc kubenswrapper[5002]: I1014 10:11:31.887487 5002 scope.go:117] "RemoveContainer" containerID="f346cc4a1dc60de31a83a0d527785353d09b17b196d072ffb455d987700d1f54" Oct 14 10:11:31 crc kubenswrapper[5002]: I1014 10:11:31.939615 5002 scope.go:117] "RemoveContainer" containerID="ef86cc8037c152771a507a4359db777da077d00262561af7c800284b74173204" Oct 14 10:11:31 crc kubenswrapper[5002]: I1014 10:11:31.997029 5002 scope.go:117] "RemoveContainer" containerID="718a01b93b7cb686b6d41dff767e8cd66e65b6de0efefc85f5c1a7267467f730" Oct 14 10:11:32 crc kubenswrapper[5002]: I1014 10:11:32.721528 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:11:32 crc kubenswrapper[5002]: E1014 10:11:32.722631 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:11:47 crc kubenswrapper[5002]: I1014 10:11:47.741644 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:11:47 crc kubenswrapper[5002]: E1014 10:11:47.742951 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:00 crc kubenswrapper[5002]: I1014 10:12:00.720989 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:12:00 crc kubenswrapper[5002]: E1014 10:12:00.722068 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:13 crc kubenswrapper[5002]: I1014 10:12:13.720724 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:12:13 crc kubenswrapper[5002]: E1014 10:12:13.721925 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:24 crc kubenswrapper[5002]: I1014 10:12:24.721027 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:12:24 crc kubenswrapper[5002]: E1014 10:12:24.722093 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:25 crc kubenswrapper[5002]: E1014 10:12:25.720975 5002 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Oct 14 10:12:32 crc kubenswrapper[5002]: I1014 10:12:32.162232 5002 scope.go:117] "RemoveContainer" containerID="8e6a7567574c02b886793172a9b80670de09b56dd581de474962d7bd184a8480" Oct 14 10:12:37 crc kubenswrapper[5002]: I1014 10:12:37.732119 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:12:37 crc kubenswrapper[5002]: E1014 10:12:37.733260 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.560027 5002 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561405 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="extract-utilities" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561439 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="extract-utilities" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561478 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561494 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561513 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561528 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561549 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="extract-content" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561563 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="extract-content" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561600 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="extract-utilities" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561618 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="extract-utilities" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561640 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="extract-content" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561658 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="extract-content" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561698 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="copy" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561714 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="copy" Oct 14 10:12:41 crc kubenswrapper[5002]: E1014 10:12:41.561737 5002 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="gather" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.561752 5002 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="gather" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.562228 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="8866cded-95f4-4436-add4-5465ad800328" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.562272 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e2c1ff-9501-434c-b3d6-97a709c93ede" containerName="registry-server" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.562305 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="copy" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.562336 5002 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eb989ee-6138-47c7-8d09-dedcf1364fd2" containerName="gather" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.566714 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.593821 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.620510 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.620572 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdvm4\" (UniqueName: \"kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.621259 5002 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.723252 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.723501 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.723565 5002 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdvm4\" (UniqueName: \"kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.723678 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.724108 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.758074 5002 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdvm4\" (UniqueName: \"kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4\") pod \"community-operators-pwhjv\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:41 crc kubenswrapper[5002]: I1014 10:12:41.899746 5002 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:42 crc kubenswrapper[5002]: I1014 10:12:42.364114 5002 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:42 crc kubenswrapper[5002]: I1014 10:12:42.759053 5002 generic.go:334] "Generic (PLEG): container finished" podID="53132ec4-1982-479c-bfed-2a768b66b1c7" containerID="c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06" exitCode=0 Oct 14 10:12:42 crc kubenswrapper[5002]: I1014 10:12:42.759126 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerDied","Data":"c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06"} Oct 14 10:12:42 crc kubenswrapper[5002]: I1014 10:12:42.759576 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerStarted","Data":"1d6d9154243e55271a191bf6f45e53ae16703436591f95bc627c848bf39a25d7"} Oct 14 10:12:43 crc kubenswrapper[5002]: I1014 10:12:43.772248 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerStarted","Data":"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f"} Oct 14 10:12:44 crc kubenswrapper[5002]: I1014 10:12:44.789795 5002 generic.go:334] "Generic (PLEG): container finished" podID="53132ec4-1982-479c-bfed-2a768b66b1c7" containerID="9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f" exitCode=0 Oct 14 10:12:44 crc kubenswrapper[5002]: I1014 10:12:44.789934 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerDied","Data":"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f"} Oct 14 10:12:45 crc kubenswrapper[5002]: I1014 10:12:45.803585 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerStarted","Data":"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3"} Oct 14 10:12:45 crc kubenswrapper[5002]: I1014 10:12:45.840538 5002 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pwhjv" podStartSLOduration=2.36879322 podStartE2EDuration="4.840514084s" podCreationTimestamp="2025-10-14 10:12:41 +0000 UTC" firstStartedPulling="2025-10-14 10:12:42.764515776 +0000 UTC m=+8495.745755228" lastFinishedPulling="2025-10-14 10:12:45.23623661 +0000 UTC m=+8498.217476092" observedRunningTime="2025-10-14 10:12:45.831892504 +0000 UTC m=+8498.813131966" watchObservedRunningTime="2025-10-14 10:12:45.840514084 +0000 UTC m=+8498.821753536" Oct 14 10:12:48 crc kubenswrapper[5002]: I1014 10:12:48.720711 5002 scope.go:117] "RemoveContainer" containerID="66b8ec6bd51249371ee60a4a960e2d06012af0aa07823967811f2b8043f248e7" Oct 14 10:12:48 crc kubenswrapper[5002]: E1014 10:12:48.721778 5002 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7btf2_openshift-machine-config-operator(6f3de631-7681-4103-8510-2b5545bbe8cd)\"" pod="openshift-machine-config-operator/machine-config-daemon-7btf2" podUID="6f3de631-7681-4103-8510-2b5545bbe8cd" Oct 14 10:12:51 crc kubenswrapper[5002]: I1014 10:12:51.900142 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:51 crc kubenswrapper[5002]: I1014 10:12:51.900569 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:52 crc kubenswrapper[5002]: I1014 10:12:52.042263 5002 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:52 crc kubenswrapper[5002]: I1014 10:12:52.957336 5002 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:53 crc kubenswrapper[5002]: I1014 10:12:53.034672 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:54 crc kubenswrapper[5002]: I1014 10:12:54.912963 5002 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pwhjv" podUID="53132ec4-1982-479c-bfed-2a768b66b1c7" containerName="registry-server" containerID="cri-o://f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3" gracePeriod=2 Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.387629 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.553197 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content\") pod \"53132ec4-1982-479c-bfed-2a768b66b1c7\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.553726 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities\") pod \"53132ec4-1982-479c-bfed-2a768b66b1c7\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.553830 5002 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdvm4\" (UniqueName: \"kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4\") pod \"53132ec4-1982-479c-bfed-2a768b66b1c7\" (UID: \"53132ec4-1982-479c-bfed-2a768b66b1c7\") " Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.554568 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities" (OuterVolumeSpecName: "utilities") pod "53132ec4-1982-479c-bfed-2a768b66b1c7" (UID: "53132ec4-1982-479c-bfed-2a768b66b1c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.562925 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4" (OuterVolumeSpecName: "kube-api-access-hdvm4") pod "53132ec4-1982-479c-bfed-2a768b66b1c7" (UID: "53132ec4-1982-479c-bfed-2a768b66b1c7"). InnerVolumeSpecName "kube-api-access-hdvm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.640185 5002 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53132ec4-1982-479c-bfed-2a768b66b1c7" (UID: "53132ec4-1982-479c-bfed-2a768b66b1c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.656616 5002 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.656651 5002 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdvm4\" (UniqueName: \"kubernetes.io/projected/53132ec4-1982-479c-bfed-2a768b66b1c7-kube-api-access-hdvm4\") on node \"crc\" DevicePath \"\"" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.656663 5002 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53132ec4-1982-479c-bfed-2a768b66b1c7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.926567 5002 generic.go:334] "Generic (PLEG): container finished" podID="53132ec4-1982-479c-bfed-2a768b66b1c7" containerID="f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3" exitCode=0 Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.926615 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerDied","Data":"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3"} Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.926651 5002 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwhjv" event={"ID":"53132ec4-1982-479c-bfed-2a768b66b1c7","Type":"ContainerDied","Data":"1d6d9154243e55271a191bf6f45e53ae16703436591f95bc627c848bf39a25d7"} Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.926671 5002 scope.go:117] "RemoveContainer" containerID="f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.926768 5002 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwhjv" Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.965338 5002 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.976326 5002 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pwhjv"] Oct 14 10:12:55 crc kubenswrapper[5002]: I1014 10:12:55.976365 5002 scope.go:117] "RemoveContainer" containerID="9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.014165 5002 scope.go:117] "RemoveContainer" containerID="c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.086294 5002 scope.go:117] "RemoveContainer" containerID="f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3" Oct 14 10:12:56 crc kubenswrapper[5002]: E1014 10:12:56.087024 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3\": container with ID starting with f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3 not found: ID does not exist" containerID="f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.087055 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3"} err="failed to get container status \"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3\": rpc error: code = NotFound desc = could not find container \"f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3\": container with ID starting with f52bdbeb1cb3074535b44188eb5d369798c66b1bac53ccdd97ede09a642595c3 not found: ID does not exist" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.087073 5002 scope.go:117] "RemoveContainer" containerID="9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f" Oct 14 10:12:56 crc kubenswrapper[5002]: E1014 10:12:56.087782 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f\": container with ID starting with 9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f not found: ID does not exist" containerID="9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.087938 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f"} err="failed to get container status \"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f\": rpc error: code = NotFound desc = could not find container \"9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f\": container with ID starting with 9512bf9b5abaa551e8979789612314f865546049b4ed92dd1a501fd4873d2b9f not found: ID does not exist" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.088024 5002 scope.go:117] "RemoveContainer" containerID="c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06" Oct 14 10:12:56 crc kubenswrapper[5002]: E1014 10:12:56.088499 5002 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06\": container with ID starting with c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06 not found: ID does not exist" containerID="c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06" Oct 14 10:12:56 crc kubenswrapper[5002]: I1014 10:12:56.088524 5002 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06"} err="failed to get container status \"c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06\": rpc error: code = NotFound desc = could not find container \"c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06\": container with ID starting with c339278b5f5a3fce4706596fe3554771e830adfd27eab3e9c27267d342a8dd06 not found: ID does not exist" Oct 14 10:12:57 crc kubenswrapper[5002]: I1014 10:12:57.736145 5002 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53132ec4-1982-479c-bfed-2a768b66b1c7" path="/var/lib/kubelet/pods/53132ec4-1982-479c-bfed-2a768b66b1c7/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073421061024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073421061017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073400034016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073400035015452 5ustar corecore